1 /* 2 * 3 * Copyright (c) 1996,1997 4 * Silicon Graphics Computer Systems, Inc. 5 * 6 * Copyright (c) 1997 7 * Moscow Center for SPARC Technology 8 * 9 * Copyright (c) 1999 10 * Boris Fomitchev 11 * 12 * This material is provided "as is", with absolutely no warranty expressed 13 * or implied. Any use is at your own risk. 14 * 15 * Permission to use or copy this software for any purpose is hereby granted 16 * without fee, provided the above notices are retained on all copies. 17 * Permission to modify the code and to distribute modified code is granted, 18 * provided the above notices are retained, and a notice that the code was 19 * modified is included with the above copyright notice. 20 * 21 */ 22 23 /* NOTE: This is an internal header file, included by other STL headers. 24 * You should not attempt to use it directly. 25 */ 26 27 // rope<_CharT,_Alloc> is a sequence of _CharT. 28 // Ropes appear to be mutable, but update operations 29 // really copy enough of the data structure to leave the original 30 // valid. Thus ropes can be logically copied by just copying 31 // a pointer value. 32 33 #ifndef _STLP_INTERNAL_ROPE_H 34 #define _STLP_INTERNAL_ROPE_H 35 36 #ifndef _STLP_INTERNAL_ALGOBASE_H 37 # include <stl/_algobase.h> 38 #endif 39 40 #if !defined (_STLP_USE_NO_IOSTREAMS) && !defined (_STLP_INTERNAL_IOSFWD) 41 # include <stl/_iosfwd.h> 42 #endif 43 44 #ifndef _STLP_INTERNAL_ALLOC_H 45 # include <stl/_alloc.h> 46 #endif 47 48 #ifndef _STLP_INTERNAL_ITERATOR_H 49 # include <stl/_iterator.h> 50 #endif 51 52 #ifndef _STLP_INTERNAL_ALGO_H 53 # include <stl/_algo.h> 54 #endif 55 56 #ifndef _STLP_INTERNAL_FUNCTION_BASE_H 57 # include <stl/_function_base.h> 58 #endif 59 60 #ifndef _STLP_INTERNAL_NUMERIC_H 61 # include <stl/_numeric.h> 62 #endif 63 64 #ifndef _STLP_INTERNAL_HASH_FUN_H 65 # include <stl/_hash_fun.h> 66 #endif 67 68 #ifndef _STLP_CHAR_TRAITS_H 69 # include <stl/char_traits.h> 70 #endif 71 72 #ifndef _STLP_INTERNAL_THREADS_H 73 # include <stl/_threads.h> 74 #endif 75 76 #ifdef _STLP_SGI_THREADS 77 # include <mutex.h> 78 #endif 79 80 #ifndef _STLP_DONT_SUPPORT_REBIND_MEMBER_TEMPLATE 81 # define _STLP_CREATE_ALLOCATOR(__atype,__a, _Tp) (_Alloc_traits<_Tp,__atype>::create_allocator(__a)) 82 #else 83 # define _STLP_CREATE_ALLOCATOR(__atype,__a, _Tp) __stl_alloc_create(__a,(_Tp*)0) 84 #endif 85 86 _STLP_BEGIN_NAMESPACE 87 88 // First a lot of forward declarations. The standard seems to require 89 // much stricter "declaration before use" than many of the implementations 90 // that preceded it. 91 template<class _CharT, _STLP_DFL_TMPL_PARAM(_Alloc, allocator<_CharT>) > class rope; 92 template<class _CharT, class _Alloc> struct _Rope_RopeConcatenation; 93 template<class _CharT, class _Alloc> struct _Rope_RopeRep; 94 template<class _CharT, class _Alloc> struct _Rope_RopeLeaf; 95 template<class _CharT, class _Alloc> struct _Rope_RopeFunction; 96 template<class _CharT, class _Alloc> struct _Rope_RopeSubstring; 97 template<class _CharT, class _Alloc> class _Rope_iterator; 98 template<class _CharT, class _Alloc> class _Rope_const_iterator; 99 template<class _CharT, class _Alloc> class _Rope_char_ref_proxy; 100 template<class _CharT, class _Alloc> class _Rope_char_ptr_proxy; 101 102 _STLP_MOVE_TO_PRIV_NAMESPACE 103 104 template <class _CharT> 105 struct _BasicCharType { typedef __false_type _Ret; }; 106 107 _STLP_TEMPLATE_NULL 108 struct _BasicCharType<char> { typedef __true_type _Ret; }; 109 110 #ifdef _STLP_HAS_WCHAR_T 111 _STLP_TEMPLATE_NULL 112 struct _BasicCharType<wchar_t> { typedef __true_type _Ret; }; 113 #endif 114 115 // Some helpers, so we can use the power algorithm on ropes. 116 // See below for why this isn't local to the implementation. 117 118 // This uses a nonstandard refcount convention. 119 // The result has refcount 0. 120 template<class _CharT, class _Alloc> 121 struct _Rope_Concat_fn 122 : public binary_function<rope<_CharT,_Alloc>, rope<_CharT,_Alloc>, 123 rope<_CharT,_Alloc> > { 124 rope<_CharT,_Alloc> operator() (const rope<_CharT,_Alloc>& __x, 125 const rope<_CharT,_Alloc>& __y) { 126 return __x + __y; 127 } 128 }; 129 130 template <class _CharT, class _Alloc> 131 inline 132 rope<_CharT,_Alloc> 133 __identity_element(_Rope_Concat_fn<_CharT, _Alloc>) 134 { return rope<_CharT,_Alloc>(); } 135 136 _STLP_MOVE_TO_STD_NAMESPACE 137 138 // Store an eos 139 template <class _CharT> 140 inline void _S_construct_null_aux(_CharT *__p, const __true_type&) 141 { *__p = 0; } 142 143 template <class _CharT> 144 inline void _S_construct_null_aux(_CharT *__p, const __false_type&) 145 { _STLP_STD::_Construct(__p); } 146 147 template <class _CharT> 148 inline void _S_construct_null(_CharT *__p) { 149 typedef typename _IsIntegral<_CharT>::_Ret _Char_Is_Integral; 150 _S_construct_null_aux(__p, _Char_Is_Integral()); 151 } 152 153 // char_producers are logically functions that generate a section of 154 // a string. These can be converted to ropes. The resulting rope 155 // invokes the char_producer on demand. This allows, for example, 156 // files to be viewed as ropes without reading the entire file. 157 template <class _CharT> 158 class char_producer { 159 public: 160 virtual ~char_producer() {} 161 virtual void operator()(size_t __start_pos, size_t __len, 162 _CharT* __buffer) = 0; 163 // Buffer should really be an arbitrary output iterator. 164 // That way we could flatten directly into an ostream, etc. 165 // This is thoroughly impossible, since iterator types don't 166 // have runtime descriptions. 167 }; 168 169 // Sequence buffers: 170 // 171 // Sequence must provide an append operation that appends an 172 // array to the sequence. Sequence buffers are useful only if 173 // appending an entire array is cheaper than appending element by element. 174 // This is true for many string representations. 175 // This should perhaps inherit from ostream<sequence::value_type> 176 // and be implemented correspondingly, so that they can be used 177 // for formatted. For the sake of portability, we don't do this yet. 178 // 179 // For now, sequence buffers behave as output iterators. But they also 180 // behave a little like basic_ostringstream<sequence::value_type> and a 181 // little like containers. 182 183 template<class _Sequence 184 # if !(defined (_STLP_NON_TYPE_TMPL_PARAM_BUG) || \ 185 defined ( _STLP_NO_DEFAULT_NON_TYPE_PARAM )) 186 , size_t _Buf_sz = 100 187 # if defined(__sgi) && !defined(__GNUC__) 188 # define __TYPEDEF_WORKAROUND 189 ,class _V = typename _Sequence::value_type 190 # endif /* __sgi */ 191 # endif /* _STLP_NON_TYPE_TMPL_PARAM_BUG */ 192 > 193 // The 3rd parameter works around a common compiler bug. 194 class sequence_buffer : public iterator <output_iterator_tag, void, void, void, void> { 195 public: 196 # ifndef __TYPEDEF_WORKAROUND 197 typedef typename _Sequence::value_type value_type; 198 typedef sequence_buffer<_Sequence 199 # if !(defined (_STLP_NON_TYPE_TMPL_PARAM_BUG) || \ 200 defined ( _STLP_NO_DEFAULT_NON_TYPE_PARAM )) 201 , _Buf_sz 202 > _Self; 203 # else /* _STLP_NON_TYPE_TMPL_PARAM_BUG */ 204 > _Self; 205 enum { _Buf_sz = 100}; 206 # endif /* _STLP_NON_TYPE_TMPL_PARAM_BUG */ 207 // # endif 208 # else /* __TYPEDEF_WORKAROUND */ 209 typedef _V value_type; 210 typedef sequence_buffer<_Sequence, _Buf_sz, _V> _Self; 211 # endif /* __TYPEDEF_WORKAROUND */ 212 protected: 213 _Sequence* _M_prefix; 214 value_type _M_buffer[_Buf_sz]; 215 size_t _M_buf_count; 216 public: 217 void flush() { 218 _M_prefix->append(_M_buffer, _M_buffer + _M_buf_count); 219 _M_buf_count = 0; 220 } 221 ~sequence_buffer() { flush(); } 222 sequence_buffer() : _M_prefix(0), _M_buf_count(0) {} 223 sequence_buffer(const _Self& __x) { 224 _M_prefix = __x._M_prefix; 225 _M_buf_count = __x._M_buf_count; 226 _STLP_STD::copy(__x._M_buffer, __x._M_buffer + __x._M_buf_count, _M_buffer); 227 } 228 sequence_buffer(_Self& __x) { 229 __x.flush(); 230 _M_prefix = __x._M_prefix; 231 _M_buf_count = 0; 232 } 233 sequence_buffer(_Sequence& __s) : _M_prefix(&__s), _M_buf_count(0) {} 234 _Self& operator= (_Self& __x) { 235 __x.flush(); 236 _M_prefix = __x._M_prefix; 237 _M_buf_count = 0; 238 return *this; 239 } 240 _Self& operator= (const _Self& __x) { 241 _M_prefix = __x._M_prefix; 242 _M_buf_count = __x._M_buf_count; 243 _STLP_STD::copy(__x._M_buffer, __x._M_buffer + __x._M_buf_count, _M_buffer); 244 return *this; 245 } 246 void push_back(value_type __x) { 247 if (_M_buf_count < _Buf_sz) { 248 _M_buffer[_M_buf_count] = __x; 249 ++_M_buf_count; 250 } else { 251 flush(); 252 _M_buffer[0] = __x; 253 _M_buf_count = 1; 254 } 255 } 256 void append(const value_type *__s, size_t __len) { 257 if (__len + _M_buf_count <= _Buf_sz) { 258 size_t __i = _M_buf_count; 259 size_t __j = 0; 260 for (; __j < __len; __i++, __j++) { 261 _M_buffer[__i] = __s[__j]; 262 } 263 _M_buf_count += __len; 264 } else if (0 == _M_buf_count) { 265 _M_prefix->append(__s, __s + __len); 266 } else { 267 flush(); 268 append(__s, __len); 269 } 270 } 271 _Self& write(const value_type *__s, size_t __len) { 272 append(__s, __len); 273 return *this; 274 } 275 _Self& put(value_type __x) { 276 push_back(__x); 277 return *this; 278 } 279 _Self& operator=(const value_type& __rhs) { 280 push_back(__rhs); 281 return *this; 282 } 283 _Self& operator*() { return *this; } 284 _Self& operator++() { return *this; } 285 _Self& operator++(int) { return *this; } 286 }; 287 288 // The following should be treated as private, at least for now. 289 template<class _CharT> 290 class _Rope_char_consumer { 291 #if !defined (_STLP_MEMBER_TEMPLATES) 292 public: 293 //Without member templates we have to use run-time parameterization. 294 // The symmetry with char_producer is accidental and temporary. 295 virtual ~_Rope_char_consumer() {} 296 virtual bool operator()(const _CharT* __buffer, size_t __len) = 0; 297 #endif 298 }; 299 300 // 301 // What follows should really be local to rope. Unfortunately, 302 // that doesn't work, since it makes it impossible to define generic 303 // equality on rope iterators. According to the draft standard, the 304 // template parameters for such an equality operator cannot be inferred 305 // from the occurence of a member class as a parameter. 306 // (SGI compilers in fact allow this, but the __result wouldn't be 307 // portable.) 308 // Similarly, some of the static member functions are member functions 309 // only to avoid polluting the global namespace, and to circumvent 310 // restrictions on type inference for template functions. 311 // 312 313 // 314 // The internal data structure for representing a rope. This is 315 // private to the implementation. A rope is really just a pointer 316 // to one of these. 317 // 318 // A few basic functions for manipulating this data structure 319 // are members of _RopeRep. Most of the more complex algorithms 320 // are implemented as rope members. 321 // 322 // Some of the static member functions of _RopeRep have identically 323 // named functions in rope that simply invoke the _RopeRep versions. 324 // 325 326 template<class _CharT, class _Alloc> 327 struct _Rope_RopeRep 328 : public _Refcount_Base 329 { 330 typedef _Rope_RopeRep<_CharT, _Alloc> _Self; 331 public: 332 // 333 // GAB: 11/09/05 334 // 335 // "__ROPE_DEPTH_SIZE" is set to one more then the "__ROPE_MAX_DEPTH". 336 // This was originally just an addition of "__ROPE_MAX_DEPTH + 1" 337 // but this addition causes the sunpro compiler to complain about 338 // multiple declarations during the initialization of "_S_min_len". 339 // Changed to be a fixed value and the sunpro compiler appears to 340 // be happy??? 341 // 342 # define __ROPE_MAX_DEPTH 45 343 # define __ROPE_DEPTH_SIZE 46 // __ROPE_MAX_DEPTH + 1 344 enum { _S_max_rope_depth = __ROPE_MAX_DEPTH }; 345 enum _Tag {_S_leaf, _S_concat, _S_substringfn, _S_function}; 346 // Apparently needed by VC++ 347 // The data fields of leaves are allocated with some 348 // extra space, to accomodate future growth and for basic 349 // character types, to hold a trailing eos character. 350 enum { _S_alloc_granularity = 8 }; 351 352 _Tag _M_tag:8; 353 bool _M_is_balanced:8; 354 355 _STLP_FORCE_ALLOCATORS(_CharT, _Alloc) 356 typedef _Alloc allocator_type; 357 358 allocator_type get_allocator() const { return allocator_type(_M_size); } 359 360 unsigned char _M_depth; 361 _CharT* _STLP_VOLATILE _M_c_string; 362 _STLP_PRIV _STLP_alloc_proxy<size_t, _CharT, allocator_type> _M_size; 363 364 #ifdef _STLP_NO_ARROW_OPERATOR 365 _Rope_RopeRep() : _Refcount_Base(1), _M_size(allocator_type(), 0) { 366 # if defined (_STLP_CHECK_RUNTIME_COMPATIBILITY) 367 _STLP_CHECK_RUNTIME_COMPATIBILITY(); 368 # endif 369 } 370 #endif 371 372 /* Flattened version of string, if needed. */ 373 /* typically 0. */ 374 /* If it's not 0, then the memory is owned */ 375 /* by this node. */ 376 /* In the case of a leaf, this may point to */ 377 /* the same memory as the data field. */ 378 _Rope_RopeRep(_Tag __t, unsigned char __d, bool __b, size_t _p_size, 379 allocator_type __a) : 380 _Refcount_Base(1), 381 _M_tag(__t), _M_is_balanced(__b), _M_depth(__d), _M_c_string(0), _M_size(__a, _p_size) { 382 #if defined (_STLP_CHECK_RUNTIME_COMPATIBILITY) 383 _STLP_CHECK_RUNTIME_COMPATIBILITY(); 384 #endif 385 } 386 387 typedef _STLP_TYPENAME _STLP_PRIV _BasicCharType<_CharT>::_Ret _IsBasicCharType; 388 389 #if 0 390 /* Please tell why this code is necessary if you uncomment it. 391 * Problem with it is that rope implementation expect that _S_rounded_up_size(n) 392 * returns a size > n in order to store the terminating null charater. When 393 * instanciation type is not a char or wchar_t this is not guaranty resulting in 394 * memory overrun. 395 */ 396 static size_t _S_rounded_up_size_aux(size_t __n, __true_type const& /*_IsBasicCharType*/) { 397 // Allow slop for in-place expansion. 398 return (__n + _S_alloc_granularity) & ~(_S_alloc_granularity - 1); 399 } 400 401 static size_t _S_rounded_up_size_aux(size_t __n, __false_type const& /*_IsBasicCharType*/) { 402 // Allow slop for in-place expansion. 403 return (__n + _S_alloc_granularity - 1) & ~(_S_alloc_granularity - 1); 404 } 405 #endif 406 // fbp : moved from RopeLeaf 407 static size_t _S_rounded_up_size(size_t __n) 408 //{ return _S_rounded_up_size_aux(__n, _IsBasicCharType()); } 409 { return (__n + _S_alloc_granularity) & ~(_S_alloc_granularity - 1); } 410 411 static void _S_free_string( _CharT* __s, size_t __len, 412 allocator_type __a) { 413 _STLP_STD::_Destroy_Range(__s, __s + __len); 414 // This has to be a static member, so this gets a bit messy 415 # ifndef _STLP_DONT_SUPPORT_REBIND_MEMBER_TEMPLATE 416 __a.deallocate(__s, _S_rounded_up_size(__len)); //*ty 03/24/2001 - restored not to use __stl_alloc_rebind() since it is not defined under _STLP_MEMBER_TEMPLATE_CLASSES 417 # else 418 __stl_alloc_rebind (__a, (_CharT*)0).deallocate(__s, _S_rounded_up_size(__len)); 419 # endif 420 } 421 422 // Deallocate data section of a leaf. 423 // This shouldn't be a member function. 424 // But its hard to do anything else at the 425 // moment, because it's templatized w.r.t. 426 // an allocator. 427 // Does nothing if __GC is defined. 428 void _M_free_c_string(); 429 void _M_free_tree(); 430 // Deallocate t. Assumes t is not 0. 431 void _M_unref_nonnil() { 432 if (_M_decr() == 0) _M_free_tree(); 433 } 434 void _M_ref_nonnil() { 435 _M_incr(); 436 } 437 static void _S_unref(_Self* __t) { 438 if (0 != __t) { 439 __t->_M_unref_nonnil(); 440 } 441 } 442 static void _S_ref(_Self* __t) { 443 if (0 != __t) __t->_M_incr(); 444 } 445 //static void _S_free_if_unref(_Self* __t) { 446 // if (0 != __t && 0 == __t->_M_ref_count) __t->_M_free_tree(); 447 //} 448 }; 449 450 template<class _CharT, class _Alloc> 451 struct _Rope_RopeLeaf : public _Rope_RopeRep<_CharT,_Alloc> { 452 public: 453 _CharT* _M_data; /* Not necessarily 0 terminated. */ 454 /* The allocated size is */ 455 /* _S_rounded_up_size(size), except */ 456 /* in the GC case, in which it */ 457 /* doesn't matter. */ 458 private: 459 typedef _Rope_RopeRep<_CharT,_Alloc> _RopeRep; 460 typedef typename _RopeRep::_IsBasicCharType _IsBasicCharType; 461 void _M_init(__true_type const& /*_IsBasicCharType*/) { 462 this->_M_c_string = _M_data; 463 } 464 void _M_init(__false_type const& /*_IsBasicCharType*/) {} 465 466 public: 467 _STLP_FORCE_ALLOCATORS(_CharT, _Alloc) 468 typedef typename _RopeRep::allocator_type allocator_type; 469 470 _Rope_RopeLeaf( _CharT* __d, size_t _p_size, allocator_type __a) 471 : _Rope_RopeRep<_CharT,_Alloc>(_RopeRep::_S_leaf, 0, true, _p_size, __a), 472 _M_data(__d) { 473 _STLP_ASSERT(_p_size > 0) 474 _M_init(_IsBasicCharType()); 475 } 476 477 # ifdef _STLP_NO_ARROW_OPERATOR 478 _Rope_RopeLeaf() {} 479 _Rope_RopeLeaf(const _Rope_RopeLeaf<_CharT, _Alloc>& ) {} 480 # endif 481 482 // The constructor assumes that d has been allocated with 483 // the proper allocator and the properly padded size. 484 // In contrast, the destructor deallocates the data: 485 ~_Rope_RopeLeaf() { 486 if (_M_data != this->_M_c_string) { 487 this->_M_free_c_string(); 488 } 489 _RopeRep::_S_free_string(_M_data, this->_M_size._M_data, this->get_allocator()); 490 } 491 }; 492 493 template<class _CharT, class _Alloc> 494 struct _Rope_RopeConcatenation : public _Rope_RopeRep<_CharT, _Alloc> { 495 private: 496 typedef _Rope_RopeRep<_CharT,_Alloc> _RopeRep; 497 498 public: 499 _RopeRep* _M_left; 500 _RopeRep* _M_right; 501 _STLP_FORCE_ALLOCATORS(_CharT, _Alloc) 502 typedef typename _RopeRep::allocator_type allocator_type; 503 _Rope_RopeConcatenation(_RopeRep* __l, _RopeRep* __r, allocator_type __a) 504 : _Rope_RopeRep<_CharT,_Alloc>(_RopeRep::_S_concat, 505 (max)(__l->_M_depth, __r->_M_depth) + 1, false, 506 __l->_M_size._M_data + __r->_M_size._M_data, __a), _M_left(__l), _M_right(__r) 507 {} 508 # ifdef _STLP_NO_ARROW_OPERATOR 509 _Rope_RopeConcatenation() {} 510 _Rope_RopeConcatenation(const _Rope_RopeConcatenation<_CharT, _Alloc>&) {} 511 # endif 512 513 ~_Rope_RopeConcatenation() { 514 this->_M_free_c_string(); 515 _M_left->_M_unref_nonnil(); 516 _M_right->_M_unref_nonnil(); 517 } 518 }; 519 520 template <class _CharT, class _Alloc> 521 struct _Rope_RopeFunction : public _Rope_RopeRep<_CharT, _Alloc> { 522 private: 523 typedef _Rope_RopeRep<_CharT,_Alloc> _RopeRep; 524 public: 525 char_producer<_CharT>* _M_fn; 526 /* 527 * Char_producer is owned by the 528 * rope and should be explicitly 529 * deleted when the rope becomes 530 * inaccessible. 531 */ 532 bool _M_delete_when_done; 533 _STLP_FORCE_ALLOCATORS(_CharT, _Alloc) 534 typedef typename _Rope_RopeRep<_CharT,_Alloc>::allocator_type allocator_type; 535 # ifdef _STLP_NO_ARROW_OPERATOR 536 _Rope_RopeFunction() {} 537 _Rope_RopeFunction(const _Rope_RopeFunction<_CharT, _Alloc>& ) {} 538 # endif 539 540 _Rope_RopeFunction(char_producer<_CharT>* __f, size_t _p_size, 541 bool __d, allocator_type __a) 542 : _Rope_RopeRep<_CharT,_Alloc>(_RopeRep::_S_function, 0, true, _p_size, __a), _M_fn(__f) 543 , _M_delete_when_done(__d) 544 { _STLP_ASSERT(_p_size > 0) } 545 546 ~_Rope_RopeFunction() { 547 this->_M_free_c_string(); 548 if (_M_delete_when_done) { 549 delete _M_fn; 550 } 551 } 552 }; 553 554 /* 555 * Substring results are usually represented using just 556 * concatenation nodes. But in the case of very long flat ropes 557 * or ropes with a functional representation that isn't practical. 558 * In that case, we represent the __result as a special case of 559 * RopeFunction, whose char_producer points back to the rope itself. 560 * In all cases except repeated substring operations and 561 * deallocation, we treat the __result as a RopeFunction. 562 */ 563 template<class _CharT, class _Alloc> 564 struct _Rope_RopeSubstring : public char_producer<_CharT>, public _Rope_RopeFunction<_CharT,_Alloc> { 565 public: 566 // XXX this whole class should be rewritten. 567 typedef _Rope_RopeRep<_CharT,_Alloc> _RopeRep; 568 _RopeRep *_M_base; // not 0 569 size_t _M_start; 570 /* virtual */ void operator()(size_t __start_pos, size_t __req_len, 571 _CharT* __buffer) { 572 typedef _Rope_RopeFunction<_CharT,_Alloc> _RopeFunction; 573 typedef _Rope_RopeLeaf<_CharT,_Alloc> _RopeLeaf; 574 switch (_M_base->_M_tag) { 575 case _RopeRep::_S_function: 576 case _RopeRep::_S_substringfn: 577 { 578 char_producer<_CharT>* __fn = 579 __STATIC_CAST(_RopeFunction*, _M_base)->_M_fn; 580 _STLP_ASSERT(__start_pos + __req_len <= this->_M_size._M_data) 581 _STLP_ASSERT(_M_start + this->_M_size._M_data <= _M_base->_M_size._M_data) 582 (*__fn)(__start_pos + _M_start, __req_len, __buffer); 583 } 584 break; 585 case _RopeRep::_S_leaf: 586 { 587 _CharT* __s = 588 __STATIC_CAST(_RopeLeaf*, _M_base)->_M_data; 589 _STLP_PRIV __ucopy_n(__s + __start_pos + _M_start, __req_len, __buffer); 590 } 591 break; 592 default: 593 _STLP_ASSERT(false) 594 ; 595 } 596 } 597 598 _STLP_FORCE_ALLOCATORS(_CharT, _Alloc) 599 typedef typename _RopeRep::allocator_type allocator_type; 600 601 _Rope_RopeSubstring(_RopeRep* __b, size_t __s, size_t __l, allocator_type __a) 602 : _Rope_RopeFunction<_CharT,_Alloc>(this, __l, false, __a), 603 _M_base(__b), _M_start(__s) { 604 _STLP_ASSERT(__l > 0) 605 _STLP_ASSERT(__s + __l <= __b->_M_size._M_data) 606 _M_base->_M_ref_nonnil(); 607 this->_M_tag = _RopeRep::_S_substringfn; 608 } 609 virtual ~_Rope_RopeSubstring() 610 { _M_base->_M_unref_nonnil(); } 611 }; 612 613 /* 614 * Self-destructing pointers to Rope_rep. 615 * These are not conventional smart pointers. Their 616 * only purpose in life is to ensure that unref is called 617 * on the pointer either at normal exit or if an exception 618 * is raised. It is the caller's responsibility to 619 * adjust reference counts when these pointers are initialized 620 * or assigned to. (This convention significantly reduces 621 * the number of potentially expensive reference count 622 * updates.) 623 */ 624 template<class _CharT, class _Alloc> 625 struct _Rope_self_destruct_ptr { 626 _Rope_RopeRep<_CharT,_Alloc>* _M_ptr; 627 ~_Rope_self_destruct_ptr() 628 { _Rope_RopeRep<_CharT,_Alloc>::_S_unref(_M_ptr); } 629 # ifdef _STLP_USE_EXCEPTIONS 630 _Rope_self_destruct_ptr() : _M_ptr(0) {} 631 # else 632 _Rope_self_destruct_ptr() {} 633 # endif 634 _Rope_self_destruct_ptr(_Rope_RopeRep<_CharT,_Alloc>* __p) : _M_ptr(__p) {} 635 _Rope_RopeRep<_CharT,_Alloc>& operator*() { return *_M_ptr; } 636 _Rope_RopeRep<_CharT,_Alloc>* operator->() { return _M_ptr; } 637 operator _Rope_RopeRep<_CharT,_Alloc>*() { return _M_ptr; } 638 _Rope_self_destruct_ptr<_CharT, _Alloc>& 639 operator= (_Rope_RopeRep<_CharT,_Alloc>* __x) 640 { _M_ptr = __x; return *this; } 641 }; 642 643 /* 644 * Dereferencing a nonconst iterator has to return something 645 * that behaves almost like a reference. It's not possible to 646 * return an actual reference since assignment requires extra 647 * work. And we would get into the same problems as with the 648 * CD2 version of basic_string. 649 */ 650 template<class _CharT, class _Alloc> 651 class _Rope_char_ref_proxy { 652 typedef _Rope_char_ref_proxy<_CharT, _Alloc> _Self; 653 friend class rope<_CharT,_Alloc>; 654 friend class _Rope_iterator<_CharT,_Alloc>; 655 friend class _Rope_char_ptr_proxy<_CharT,_Alloc>; 656 typedef _Rope_self_destruct_ptr<_CharT,_Alloc> _Self_destruct_ptr; 657 typedef _Rope_RopeRep<_CharT,_Alloc> _RopeRep; 658 typedef rope<_CharT,_Alloc> _My_rope; 659 size_t _M_pos; 660 _CharT _M_current; 661 bool _M_current_valid; 662 _My_rope* _M_root; // The whole rope. 663 public: 664 _Rope_char_ref_proxy(_My_rope* __r, size_t __p) : 665 _M_pos(__p), _M_current_valid(false), _M_root(__r) {} 666 _Rope_char_ref_proxy(const _Self& __x) : 667 _M_pos(__x._M_pos), _M_current_valid(false), _M_root(__x._M_root) {} 668 // Don't preserve cache if the reference can outlive the 669 // expression. We claim that's not possible without calling 670 // a copy constructor or generating reference to a proxy 671 // reference. We declare the latter to have undefined semantics. 672 _Rope_char_ref_proxy(_My_rope* __r, size_t __p, _CharT __c) 673 : _M_pos(__p), _M_current(__c), _M_current_valid(true), _M_root(__r) {} 674 inline operator _CharT () const; 675 _Self& operator= (_CharT __c); 676 _Rope_char_ptr_proxy<_CharT, _Alloc> operator& () const; 677 _Self& operator= (const _Self& __c) { 678 return operator=((_CharT)__c); 679 } 680 }; 681 682 #ifdef _STLP_FUNCTION_TMPL_PARTIAL_ORDER 683 template<class _CharT, class __Alloc> 684 inline void swap(_Rope_char_ref_proxy <_CharT, __Alloc > __a, 685 _Rope_char_ref_proxy <_CharT, __Alloc > __b) { 686 _CharT __tmp = __a; 687 __a = __b; 688 __b = __tmp; 689 } 690 #else 691 // There is no really acceptable way to handle this. The default 692 // definition of swap doesn't work for proxy references. 693 // It can't really be made to work, even with ugly hacks, since 694 // the only unusual operation it uses is the copy constructor, which 695 // is needed for other purposes. We provide a macro for 696 // full specializations, and instantiate the most common case. 697 # define _ROPE_SWAP_SPECIALIZATION(_CharT, __Alloc) \ 698 inline void swap(_Rope_char_ref_proxy <_CharT, __Alloc > __a, \ 699 _Rope_char_ref_proxy <_CharT, __Alloc > __b) { \ 700 _CharT __tmp = __a; \ 701 __a = __b; \ 702 __b = __tmp; \ 703 } 704 705 _ROPE_SWAP_SPECIALIZATION(char, allocator<char>) 706 707 # ifndef _STLP_NO_WCHAR_T 708 _ROPE_SWAP_SPECIALIZATION(wchar_t, allocator<wchar_t>) 709 # endif 710 711 #endif /* !_STLP_FUNCTION_TMPL_PARTIAL_ORDER */ 712 713 template<class _CharT, class _Alloc> 714 class _Rope_char_ptr_proxy { 715 // XXX this class should be rewritten. 716 public: 717 typedef _Rope_char_ptr_proxy<_CharT, _Alloc> _Self; 718 friend class _Rope_char_ref_proxy<_CharT,_Alloc>; 719 size_t _M_pos; 720 rope<_CharT,_Alloc>* _M_root; // The whole rope. 721 722 _Rope_char_ptr_proxy(const _Rope_char_ref_proxy<_CharT,_Alloc>& __x) 723 : _M_pos(__x._M_pos), _M_root(__x._M_root) {} 724 _Rope_char_ptr_proxy(const _Self& __x) 725 : _M_pos(__x._M_pos), _M_root(__x._M_root) {} 726 _Rope_char_ptr_proxy() {} 727 _Rope_char_ptr_proxy(_CharT* __x) : _M_pos(0), _M_root(0) { 728 _STLP_ASSERT(0 == __x) 729 } 730 _Self& operator= (const _Self& __x) { 731 _M_pos = __x._M_pos; 732 _M_root = __x._M_root; 733 return *this; 734 } 735 736 _Rope_char_ref_proxy<_CharT,_Alloc> operator*() const { 737 return _Rope_char_ref_proxy<_CharT,_Alloc>(_M_root, _M_pos); 738 } 739 }; 740 741 742 /* 743 * Rope iterators: 744 * Unlike in the C version, we cache only part of the stack 745 * for rope iterators, since they must be efficiently copyable. 746 * When we run out of cache, we have to reconstruct the iterator 747 * value. 748 * Pointers from iterators are not included in reference counts. 749 * Iterators are assumed to be thread private. Ropes can 750 * be shared. 751 */ 752 template<class _CharT, class _Alloc> 753 class _Rope_iterator_base 754 /* : public random_access_iterator<_CharT, ptrdiff_t> */ 755 { 756 friend class rope<_CharT,_Alloc>; 757 typedef _Rope_iterator_base<_CharT, _Alloc> _Self; 758 typedef _Rope_RopeConcatenation<_CharT,_Alloc> _RopeConcat; 759 public: 760 typedef _Rope_RopeRep<_CharT,_Alloc> _RopeRep; 761 762 enum { _S_path_cache_len = 4 }; // Must be <= 9 because of _M_path_direction. 763 enum { _S_iterator_buf_len = 15 }; 764 size_t _M_current_pos; 765 // The whole rope. 766 _RopeRep* _M_root; 767 // Starting position for current leaf 768 size_t _M_leaf_pos; 769 // Buffer possibly containing current char. 770 _CharT* _M_buf_start; 771 // Pointer to current char in buffer, != 0 ==> buffer valid. 772 _CharT* _M_buf_ptr; 773 // One past __last valid char in buffer. 774 _CharT* _M_buf_end; 775 776 // What follows is the path cache. We go out of our 777 // way to make this compact. 778 // Path_end contains the bottom section of the path from 779 // the root to the current leaf. 780 struct { 781 # if defined (__BORLANDC__) && (__BORLANDC__ < 0x560) 782 _RopeRep const*_M_data[4]; 783 # else 784 _RopeRep const*_M_data[_S_path_cache_len]; 785 # endif 786 } _M_path_end; 787 // Last valid __pos in path_end; 788 // _M_path_end[0] ... _M_path_end[_M_leaf_index-1] 789 // point to concatenation nodes. 790 int _M_leaf_index; 791 // (_M_path_directions >> __i) & 1 is 1 792 // if we got from _M_path_end[leaf_index - __i - 1] 793 // to _M_path_end[leaf_index - __i] by going to the 794 // __right. Assumes path_cache_len <= 9. 795 unsigned char _M_path_directions; 796 // Short buffer for surrounding chars. 797 // This is useful primarily for 798 // RopeFunctions. We put the buffer 799 // here to avoid locking in the 800 // multithreaded case. 801 // The cached path is generally assumed to be valid 802 // only if the buffer is valid. 803 struct { 804 # if defined (__BORLANDC__) && (__BORLANDC__ < 0x560) 805 _CharT _M_data[15]; 806 # else 807 _CharT _M_data[_S_iterator_buf_len]; 808 # endif 809 } _M_tmp_buf; 810 811 // Set buffer contents given path cache. 812 static void _S_setbuf(_Rope_iterator_base<_CharT, _Alloc>& __x); 813 // Set buffer contents and path cache. 814 static void _S_setcache(_Rope_iterator_base<_CharT, _Alloc>& __x); 815 // As above, but assumes path cache is valid for previous posn. 816 static void _S_setcache_for_incr(_Rope_iterator_base<_CharT, _Alloc>& __x); 817 _Rope_iterator_base() {} 818 _Rope_iterator_base(_RopeRep* __root, size_t __pos) 819 : _M_current_pos(__pos),_M_root(__root), _M_buf_ptr(0) {} 820 void _M_incr(size_t __n); 821 void _M_decr(size_t __n); 822 public: 823 size_t index() const { return _M_current_pos; } 824 private: 825 void _M_copy_buf(const _Self& __x) { 826 _M_tmp_buf = __x._M_tmp_buf; 827 if (__x._M_buf_start == __x._M_tmp_buf._M_data) { 828 _M_buf_start = _M_tmp_buf._M_data; 829 _M_buf_end = _M_buf_start + (__x._M_buf_end - __x._M_buf_start); 830 _M_buf_ptr = _M_buf_start + (__x._M_buf_ptr - __x._M_buf_start); 831 } else { 832 _M_buf_end = __x._M_buf_end; 833 } 834 } 835 836 public: 837 _Rope_iterator_base(const _Self& __x) : 838 _M_current_pos(__x._M_current_pos), 839 _M_root(__x._M_root), 840 _M_leaf_pos( __x._M_leaf_pos ), 841 _M_buf_start(__x._M_buf_start), 842 _M_buf_ptr(__x._M_buf_ptr), 843 _M_path_end(__x._M_path_end), 844 _M_leaf_index(__x._M_leaf_index), 845 _M_path_directions(__x._M_path_directions) 846 { 847 if (0 != __x._M_buf_ptr) { 848 _M_copy_buf(__x); 849 } 850 } 851 _Self& operator = (const _Self& __x) 852 { 853 _M_current_pos = __x._M_current_pos; 854 _M_root = __x._M_root; 855 _M_buf_start = __x._M_buf_start; 856 _M_buf_ptr = __x._M_buf_ptr; 857 _M_path_end = __x._M_path_end; 858 _M_leaf_index = __x._M_leaf_index; 859 _M_path_directions = __x._M_path_directions; 860 _M_leaf_pos = __x._M_leaf_pos; 861 if (0 != __x._M_buf_ptr) { 862 _M_copy_buf(__x); 863 } 864 return *this; 865 } 866 }; 867 868 template<class _CharT, class _Alloc> class _Rope_iterator; 869 870 template<class _CharT, class _Alloc> 871 class _Rope_const_iterator : public _Rope_iterator_base<_CharT,_Alloc> { 872 friend class rope<_CharT,_Alloc>; 873 typedef _Rope_const_iterator<_CharT, _Alloc> _Self; 874 typedef _Rope_iterator_base<_CharT,_Alloc> _Base; 875 // protected: 876 public: 877 # ifndef _STLP_HAS_NO_NAMESPACES 878 typedef _Rope_RopeRep<_CharT,_Alloc> _RopeRep; 879 // The one from the base class may not be directly visible. 880 # endif 881 _Rope_const_iterator(const _RopeRep* __root, size_t __pos): 882 _Rope_iterator_base<_CharT,_Alloc>(__CONST_CAST(_RopeRep*,__root), __pos) 883 // Only nonconst iterators modify root ref count 884 {} 885 public: 886 typedef _CharT reference; // Really a value. Returning a reference 887 // Would be a mess, since it would have 888 // to be included in refcount. 889 typedef const _CharT* pointer; 890 typedef _CharT value_type; 891 typedef ptrdiff_t difference_type; 892 typedef random_access_iterator_tag iterator_category; 893 894 public: 895 _Rope_const_iterator() {} 896 _Rope_const_iterator(const _Self& __x) : 897 _Rope_iterator_base<_CharT,_Alloc>(__x) { } 898 _Rope_const_iterator(const _Rope_iterator<_CharT,_Alloc>& __x): 899 _Rope_iterator_base<_CharT,_Alloc>(__x) {} 900 _Rope_const_iterator(const rope<_CharT,_Alloc>& __r, size_t __pos) : 901 _Rope_iterator_base<_CharT,_Alloc>(__r._M_tree_ptr._M_data, __pos) {} 902 _Self& operator= (const _Self& __x) { 903 _Base::operator=(__x); 904 return *this; 905 } 906 reference operator*() { 907 if (0 == this->_M_buf_ptr) 908 #if !defined (__DMC__) 909 _S_setcache(*this); 910 #else 911 { _Rope_iterator_base<_CharT, _Alloc>* __x = this; _S_setcache(*__x); } 912 #endif 913 return *(this->_M_buf_ptr); 914 } 915 _Self& operator++() 916 { 917 if ( this->_M_buf_ptr != 0 ) { 918 _CharT *__next = this->_M_buf_ptr + 1; 919 if ( __next < this->_M_buf_end ) { 920 this->_M_buf_ptr = __next; 921 ++this->_M_current_pos; 922 return *this; 923 } 924 } 925 this->_M_incr(1); 926 return *this; 927 } 928 _Self& operator+=(ptrdiff_t __n) { 929 if (__n >= 0) { 930 this->_M_incr(__n); 931 } else { 932 this->_M_decr(-__n); 933 } 934 return *this; 935 } 936 _Self& operator--() { 937 this->_M_decr(1); 938 return *this; 939 } 940 _Self& operator-=(ptrdiff_t __n) { 941 if (__n >= 0) { 942 this->_M_decr(__n); 943 } else { 944 this->_M_incr(-__n); 945 } 946 return *this; 947 } 948 _Self operator++(int) { 949 size_t __old_pos = this->_M_current_pos; 950 this->_M_incr(1); 951 return _Rope_const_iterator<_CharT,_Alloc>(this->_M_root, __old_pos); 952 // This makes a subsequent dereference expensive. 953 // Perhaps we should instead copy the iterator 954 // if it has a valid cache? 955 } 956 _Self operator--(int) { 957 size_t __old_pos = this->_M_current_pos; 958 this->_M_decr(1); 959 return _Rope_const_iterator<_CharT,_Alloc>(this->_M_root, __old_pos); 960 } 961 inline reference operator[](size_t __n); 962 }; 963 964 template<class _CharT, class _Alloc> 965 class _Rope_iterator : public _Rope_iterator_base<_CharT,_Alloc> { 966 friend class rope<_CharT,_Alloc>; 967 typedef _Rope_iterator<_CharT, _Alloc> _Self; 968 typedef _Rope_iterator_base<_CharT,_Alloc> _Base; 969 typedef _Rope_RopeRep<_CharT,_Alloc> _RopeRep; 970 971 public: 972 rope<_CharT,_Alloc>* _M_root_rope; 973 // root is treated as a cached version of this, 974 // and is used to detect changes to the underlying 975 // rope. 976 // Root is included in the reference count. 977 // This is necessary so that we can detect changes reliably. 978 // Unfortunately, it requires careful bookkeeping for the 979 // nonGC case. 980 _Rope_iterator(rope<_CharT,_Alloc>* __r, size_t __pos); 981 982 void _M_check(); 983 public: 984 typedef _Rope_char_ref_proxy<_CharT,_Alloc> reference; 985 typedef _Rope_char_ref_proxy<_CharT,_Alloc>* pointer; 986 typedef _CharT value_type; 987 typedef ptrdiff_t difference_type; 988 typedef random_access_iterator_tag iterator_category; 989 public: 990 ~_Rope_iterator() { //*TY 5/6/00 - added dtor to balance reference count 991 _RopeRep::_S_unref(this->_M_root); 992 } 993 994 rope<_CharT,_Alloc>& container() { return *_M_root_rope; } 995 _Rope_iterator() { 996 this->_M_root = 0; // Needed for reference counting. 997 } 998 _Rope_iterator(const _Self& __x) : 999 _Rope_iterator_base<_CharT,_Alloc>(__x) { 1000 _M_root_rope = __x._M_root_rope; 1001 _RopeRep::_S_ref(this->_M_root); 1002 } 1003 _Rope_iterator(rope<_CharT,_Alloc>& __r, size_t __pos); 1004 _Self& operator= (const _Self& __x) { 1005 _RopeRep* __old = this->_M_root; 1006 _RopeRep::_S_ref(__x._M_root); 1007 _Base::operator=(__x); 1008 _M_root_rope = __x._M_root_rope; 1009 _RopeRep::_S_unref(__old); 1010 return *this; 1011 } 1012 reference operator*() { 1013 _M_check(); 1014 if (0 == this->_M_buf_ptr) { 1015 return reference(_M_root_rope, this->_M_current_pos); 1016 } else { 1017 return reference(_M_root_rope, this->_M_current_pos, *(this->_M_buf_ptr)); 1018 } 1019 } 1020 _Self& operator++() { 1021 this->_M_incr(1); 1022 return *this; 1023 } 1024 _Self& operator+=(ptrdiff_t __n) { 1025 if (__n >= 0) { 1026 this->_M_incr(__n); 1027 } else { 1028 this->_M_decr(-__n); 1029 } 1030 return *this; 1031 } 1032 _Self& operator--() { 1033 this->_M_decr(1); 1034 return *this; 1035 } 1036 _Self& operator-=(ptrdiff_t __n) { 1037 if (__n >= 0) { 1038 this->_M_decr(__n); 1039 } else { 1040 this->_M_incr(-__n); 1041 } 1042 return *this; 1043 } 1044 _Self operator++(int) { 1045 size_t __old_pos = this->_M_current_pos; 1046 this->_M_incr(1); 1047 return _Self(_M_root_rope, __old_pos); 1048 } 1049 _Self operator--(int) { 1050 size_t __old_pos = this->_M_current_pos; 1051 this->_M_decr(1); 1052 return _Self(_M_root_rope, __old_pos); 1053 } 1054 reference operator[](ptrdiff_t __n) { 1055 return reference(_M_root_rope, this->_M_current_pos + __n); 1056 } 1057 }; 1058 1059 # ifdef _STLP_USE_OLD_HP_ITERATOR_QUERIES 1060 template <class _CharT, class _Alloc> 1061 inline random_access_iterator_tag 1062 iterator_category(const _Rope_iterator<_CharT,_Alloc>&) { return random_access_iterator_tag();} 1063 template <class _CharT, class _Alloc> 1064 inline _CharT* value_type(const _Rope_iterator<_CharT,_Alloc>&) { return 0; } 1065 template <class _CharT, class _Alloc> 1066 inline ptrdiff_t* distance_type(const _Rope_iterator<_CharT,_Alloc>&) { return 0; } 1067 template <class _CharT, class _Alloc> 1068 inline random_access_iterator_tag 1069 iterator_category(const _Rope_const_iterator<_CharT,_Alloc>&) { return random_access_iterator_tag(); } 1070 template <class _CharT, class _Alloc> 1071 inline _CharT* value_type(const _Rope_const_iterator<_CharT,_Alloc>&) { return 0; } 1072 template <class _CharT, class _Alloc> 1073 inline ptrdiff_t* distance_type(const _Rope_const_iterator<_CharT,_Alloc>&) { return 0; } 1074 #endif /* _STLP_USE_OLD_HP_ITERATOR_QUERIES */ 1075 1076 template <class _CharT, class _Alloc, class _CharConsumer> 1077 bool _S_apply_to_pieces(_CharConsumer& __c, 1078 _Rope_RopeRep<_CharT, _Alloc> *__r, 1079 size_t __begin, size_t __end); 1080 // begin and end are assumed to be in range. 1081 1082 template <class _CharT, class _Alloc> 1083 class rope 1084 #if defined (_STLP_USE_PARTIAL_SPEC_WORKAROUND) 1085 : public __stlport_class<rope<_CharT, _Alloc> > 1086 #endif 1087 { 1088 typedef rope<_CharT,_Alloc> _Self; 1089 public: 1090 typedef _CharT value_type; 1091 typedef ptrdiff_t difference_type; 1092 typedef size_t size_type; 1093 typedef _CharT const_reference; 1094 typedef const _CharT* const_pointer; 1095 typedef _Rope_iterator<_CharT,_Alloc> iterator; 1096 typedef _Rope_const_iterator<_CharT,_Alloc> const_iterator; 1097 typedef _Rope_char_ref_proxy<_CharT,_Alloc> reference; 1098 typedef _Rope_char_ptr_proxy<_CharT,_Alloc> pointer; 1099 1100 friend class _Rope_iterator<_CharT,_Alloc>; 1101 friend class _Rope_const_iterator<_CharT,_Alloc>; 1102 friend struct _Rope_RopeRep<_CharT,_Alloc>; 1103 friend class _Rope_iterator_base<_CharT,_Alloc>; 1104 friend class _Rope_char_ptr_proxy<_CharT,_Alloc>; 1105 friend class _Rope_char_ref_proxy<_CharT,_Alloc>; 1106 friend struct _Rope_RopeSubstring<_CharT,_Alloc>; 1107 1108 _STLP_DECLARE_RANDOM_ACCESS_REVERSE_ITERATORS; 1109 1110 protected: 1111 typedef _CharT* _Cstrptr; 1112 1113 static _CharT _S_empty_c_str[1]; 1114 1115 enum { _S_copy_max = 23 }; 1116 // For strings shorter than _S_copy_max, we copy to 1117 // concatenate. 1118 1119 typedef _Rope_RopeRep<_CharT, _Alloc> _RopeRep; 1120 typedef typename _RopeRep::_IsBasicCharType _IsBasicCharType; 1121 1122 public: 1123 _STLP_FORCE_ALLOCATORS(_CharT, _Alloc) 1124 typedef _Alloc allocator_type; 1125 1126 public: 1127 // The only data member of a rope: 1128 _STLP_PRIV _STLP_alloc_proxy<_RopeRep*, _CharT, allocator_type> _M_tree_ptr; 1129 1130 public: 1131 allocator_type get_allocator() const { return allocator_type(_M_tree_ptr); } 1132 1133 public: 1134 typedef _Rope_RopeConcatenation<_CharT,_Alloc> _RopeConcatenation; 1135 typedef _Rope_RopeLeaf<_CharT,_Alloc> _RopeLeaf; 1136 typedef _Rope_RopeFunction<_CharT,_Alloc> _RopeFunction; 1137 typedef _Rope_RopeSubstring<_CharT,_Alloc> _RopeSubstring; 1138 1139 // Retrieve a character at the indicated position. 1140 static _CharT _S_fetch(_RopeRep* __r, size_type __pos); 1141 1142 // Obtain a pointer to the character at the indicated position. 1143 // The pointer can be used to change the character. 1144 // If such a pointer cannot be produced, as is frequently the 1145 // case, 0 is returned instead. 1146 // (Returns nonzero only if all nodes in the path have a refcount 1147 // of 1.) 1148 static _CharT* _S_fetch_ptr(_RopeRep* __r, size_type __pos); 1149 1150 static void _S_unref(_RopeRep* __t) { 1151 _RopeRep::_S_unref(__t); 1152 } 1153 static void _S_ref(_RopeRep* __t) { 1154 _RopeRep::_S_ref(__t); 1155 } 1156 1157 typedef _Rope_self_destruct_ptr<_CharT,_Alloc> _Self_destruct_ptr; 1158 1159 // _Result is counted in refcount. 1160 static _RopeRep* _S_substring(_RopeRep* __base, 1161 size_t __start, size_t __endp1); 1162 1163 static _RopeRep* _S_concat_char_iter(_RopeRep* __r, 1164 const _CharT* __iter, size_t __slen); 1165 // Concatenate rope and char ptr, copying __s. 1166 // Should really take an arbitrary iterator. 1167 // Result is counted in refcount. 1168 static _RopeRep* _S_destr_concat_char_iter(_RopeRep* __r, 1169 const _CharT* __iter, size_t __slen); 1170 // As above, but one reference to __r is about to be 1171 // destroyed. Thus the pieces may be recycled if all 1172 // relevent reference counts are 1. 1173 1174 // General concatenation on _RopeRep. _Result 1175 // has refcount of 1. Adjusts argument refcounts. 1176 static _RopeRep* _S_concat_rep(_RopeRep* __left, _RopeRep* __right); 1177 1178 public: 1179 #if defined (_STLP_MEMBER_TEMPLATES) 1180 template <class _CharConsumer> 1181 #else 1182 typedef _Rope_char_consumer<_CharT> _CharConsumer; 1183 #endif 1184 void apply_to_pieces(size_t __begin, size_t __end, 1185 _CharConsumer& __c) const 1186 { _S_apply_to_pieces(__c, _M_tree_ptr._M_data, __begin, __end); } 1187 1188 protected: 1189 1190 static size_t _S_rounded_up_size(size_t __n) 1191 { return _RopeRep::_S_rounded_up_size(__n); } 1192 1193 // Allocate and construct a RopeLeaf using the supplied allocator 1194 // Takes ownership of s instead of copying. 1195 static _RopeLeaf* _S_new_RopeLeaf(_CharT *__s, 1196 size_t _p_size, allocator_type __a) { 1197 _RopeLeaf* __space = _STLP_CREATE_ALLOCATOR(allocator_type, __a, 1198 _RopeLeaf).allocate(1); 1199 _STLP_TRY { 1200 new(__space) _RopeLeaf(__s, _p_size, __a); 1201 } 1202 _STLP_UNWIND(_STLP_CREATE_ALLOCATOR(allocator_type,__a, 1203 _RopeLeaf).deallocate(__space, 1)) 1204 return __space; 1205 } 1206 1207 static _RopeConcatenation* _S_new_RopeConcatenation(_RopeRep* __left, _RopeRep* __right, 1208 allocator_type __a) { 1209 _RopeConcatenation* __space = _STLP_CREATE_ALLOCATOR(allocator_type, __a, 1210 _RopeConcatenation).allocate(1); 1211 return new(__space) _RopeConcatenation(__left, __right, __a); 1212 } 1213 1214 static _RopeFunction* _S_new_RopeFunction(char_producer<_CharT>* __f, 1215 size_t _p_size, bool __d, allocator_type __a) { 1216 _RopeFunction* __space = _STLP_CREATE_ALLOCATOR(allocator_type, __a, 1217 _RopeFunction).allocate(1); 1218 return new(__space) _RopeFunction(__f, _p_size, __d, __a); 1219 } 1220 1221 static _RopeSubstring* _S_new_RopeSubstring(_Rope_RopeRep<_CharT,_Alloc>* __b, size_t __s, 1222 size_t __l, allocator_type __a) { 1223 _RopeSubstring* __space = _STLP_CREATE_ALLOCATOR(allocator_type, __a, 1224 _RopeSubstring).allocate(1); 1225 return new(__space) _RopeSubstring(__b, __s, __l, __a); 1226 } 1227 1228 static 1229 _RopeLeaf* _S_RopeLeaf_from_unowned_char_ptr(const _CharT *__s, 1230 size_t _p_size, allocator_type __a) { 1231 if (0 == _p_size) return 0; 1232 1233 _CharT* __buf = _STLP_CREATE_ALLOCATOR(allocator_type,__a, _CharT).allocate(_S_rounded_up_size(_p_size)); 1234 1235 _STLP_PRIV __ucopy_n(__s, _p_size, __buf); 1236 _S_construct_null(__buf + _p_size); 1237 1238 _STLP_TRY { 1239 return _S_new_RopeLeaf(__buf, _p_size, __a); 1240 } 1241 _STLP_UNWIND(_RopeRep::_S_free_string(__buf, _p_size, __a)) 1242 _STLP_RET_AFTER_THROW(0) 1243 } 1244 1245 1246 // Concatenation of nonempty strings. 1247 // Always builds a concatenation node. 1248 // Rebalances if the result is too deep. 1249 // Result has refcount 1. 1250 // Does not increment left and right ref counts even though 1251 // they are referenced. 1252 static _RopeRep* 1253 _S_tree_concat(_RopeRep* __left, _RopeRep* __right); 1254 1255 // Concatenation helper functions 1256 static _RopeLeaf* 1257 _S_leaf_concat_char_iter(_RopeLeaf* __r, 1258 const _CharT* __iter, size_t __slen); 1259 // Concatenate by copying leaf. 1260 // should take an arbitrary iterator 1261 // result has refcount 1. 1262 static _RopeLeaf* _S_destr_leaf_concat_char_iter 1263 (_RopeLeaf* __r, const _CharT* __iter, size_t __slen); 1264 // A version that potentially clobbers __r if __r->_M_ref_count == 1. 1265 1266 1267 // A helper function for exponentiating strings. 1268 // This uses a nonstandard refcount convention. 1269 // The result has refcount 0. 1270 typedef _STLP_PRIV _Rope_Concat_fn<_CharT,_Alloc> _Concat_fn; 1271 #if !defined (__GNUC__) || (__GNUC__ < 3) 1272 friend _Concat_fn; 1273 #else 1274 friend struct _STLP_PRIV _Rope_Concat_fn<_CharT,_Alloc>; 1275 #endif 1276 1277 public: 1278 static size_t _S_char_ptr_len(const _CharT* __s) { 1279 return char_traits<_CharT>::length(__s); 1280 } 1281 1282 public: /* for operators */ 1283 rope(_RopeRep* __t, const allocator_type& __a = allocator_type()) 1284 : _M_tree_ptr(__a, __t) { } 1285 private: 1286 // Copy __r to the _CharT buffer. 1287 // Returns __buffer + __r->_M_size._M_data. 1288 // Assumes that buffer is uninitialized. 1289 static _CharT* _S_flatten(_RopeRep* __r, _CharT* __buffer); 1290 1291 // Again, with explicit starting position and length. 1292 // Assumes that buffer is uninitialized. 1293 static _CharT* _S_flatten(_RopeRep* __r, 1294 size_t __start, size_t __len, 1295 _CharT* __buffer); 1296 1297 // fbp : HP aCC prohibits access to protected min_len from within static methods ( ?? ) 1298 public: 1299 static const unsigned long _S_min_len[__ROPE_DEPTH_SIZE]; 1300 protected: 1301 static bool _S_is_balanced(_RopeRep* __r) 1302 { return (__r->_M_size._M_data >= _S_min_len[__r->_M_depth]); } 1303 1304 static bool _S_is_almost_balanced(_RopeRep* __r) { 1305 return (__r->_M_depth == 0 || 1306 __r->_M_size._M_data >= _S_min_len[__r->_M_depth - 1]); 1307 } 1308 1309 static bool _S_is_roughly_balanced(_RopeRep* __r) { 1310 return (__r->_M_depth <= 1 || 1311 __r->_M_size._M_data >= _S_min_len[__r->_M_depth - 2]); 1312 } 1313 1314 // Assumes the result is not empty. 1315 static _RopeRep* _S_concat_and_set_balanced(_RopeRep* __left, 1316 _RopeRep* __right) { 1317 _RopeRep* __result = _S_concat_rep(__left, __right); 1318 if (_S_is_balanced(__result)) __result->_M_is_balanced = true; 1319 return __result; 1320 } 1321 1322 // The basic rebalancing operation. Logically copies the 1323 // rope. The result has refcount of 1. The client will 1324 // usually decrement the reference count of __r. 1325 // The result is within height 2 of balanced by the above 1326 // definition. 1327 static _RopeRep* _S_balance(_RopeRep* __r); 1328 1329 // Add all unbalanced subtrees to the forest of balanceed trees. 1330 // Used only by balance. 1331 static void _S_add_to_forest(_RopeRep*__r, _RopeRep** __forest); 1332 1333 // Add __r to forest, assuming __r is already balanced. 1334 static void _S_add_leaf_to_forest(_RopeRep* __r, _RopeRep** __forest); 1335 1336 #ifdef _STLP_DEBUG 1337 // Print to stdout, exposing structure 1338 static void _S_dump(_RopeRep* __r, int __indent = 0); 1339 #endif 1340 1341 // Return -1, 0, or 1 if __x < __y, __x == __y, or __x > __y resp. 1342 static int _S_compare(const _RopeRep* __x, const _RopeRep* __y); 1343 1344 void _STLP_FUNCTION_THROWS _M_throw_out_of_range() const; 1345 1346 void _M_reset(_RopeRep* __r) { 1347 //if (__r != _M_tree_ptr._M_data) { 1348 _S_unref(_M_tree_ptr._M_data); 1349 _M_tree_ptr._M_data = __r; 1350 //} 1351 } 1352 1353 public: 1354 bool empty() const { return 0 == _M_tree_ptr._M_data; } 1355 1356 // Comparison member function. This is public only for those 1357 // clients that need a ternary comparison. Others 1358 // should use the comparison operators below. 1359 int compare(const _Self& __y) const { 1360 return _S_compare(_M_tree_ptr._M_data, __y._M_tree_ptr._M_data); 1361 } 1362 1363 rope(const _CharT* __s, const allocator_type& __a = allocator_type()) 1364 : _M_tree_ptr(__a, _S_RopeLeaf_from_unowned_char_ptr(__s, _S_char_ptr_len(__s),__a)) 1365 {} 1366 1367 rope(const _CharT* __s, size_t __len, 1368 const allocator_type& __a = allocator_type()) 1369 : _M_tree_ptr(__a, (_S_RopeLeaf_from_unowned_char_ptr(__s, __len, __a))) 1370 {} 1371 1372 // Should perhaps be templatized with respect to the iterator type 1373 // and use Sequence_buffer. (It should perhaps use sequence_buffer 1374 // even now.) 1375 rope(const _CharT *__s, const _CharT *__e, 1376 const allocator_type& __a = allocator_type()) 1377 : _M_tree_ptr(__a, _S_RopeLeaf_from_unowned_char_ptr(__s, __e - __s, __a)) 1378 {} 1379 1380 rope(const const_iterator& __s, const const_iterator& __e, 1381 const allocator_type& __a = allocator_type()) 1382 : _M_tree_ptr(__a, _S_substring(__s._M_root, __s._M_current_pos, 1383 __e._M_current_pos)) 1384 {} 1385 1386 rope(const iterator& __s, const iterator& __e, 1387 const allocator_type& __a = allocator_type()) 1388 : _M_tree_ptr(__a, _S_substring(__s._M_root, __s._M_current_pos, 1389 __e._M_current_pos)) 1390 {} 1391 1392 rope(_CharT __c, const allocator_type& __a = allocator_type()) 1393 : _M_tree_ptr(__a, (_RopeRep*)0) { 1394 _CharT* __buf = _M_tree_ptr.allocate(_S_rounded_up_size(1)); 1395 1396 _Copy_Construct(__buf, __c); 1397 _S_construct_null(__buf + 1); 1398 1399 _STLP_TRY { 1400 _M_tree_ptr._M_data = _S_new_RopeLeaf(__buf, 1, __a); 1401 } 1402 _STLP_UNWIND(_RopeRep::_S_free_string(__buf, 1, __a)) 1403 } 1404 1405 rope(size_t __n, _CharT __c, 1406 const allocator_type& __a = allocator_type()): 1407 _M_tree_ptr(__a, (_RopeRep*)0) { 1408 if (0 == __n) 1409 return; 1410 1411 rope<_CharT,_Alloc> __result; 1412 # define __exponentiate_threshold size_t(32) 1413 _RopeRep* __remainder; 1414 rope<_CharT,_Alloc> __remainder_rope; 1415 1416 // gcc-2.7.2 bugs 1417 typedef _STLP_PRIV _Rope_Concat_fn<_CharT,_Alloc> _Concat_fn; 1418 1419 size_t __exponent = __n / __exponentiate_threshold; 1420 size_t __rest = __n % __exponentiate_threshold; 1421 if (0 == __rest) { 1422 __remainder = 0; 1423 } else { 1424 _CharT* __rest_buffer = _M_tree_ptr.allocate(_S_rounded_up_size(__rest)); 1425 uninitialized_fill_n(__rest_buffer, __rest, __c); 1426 _S_construct_null(__rest_buffer + __rest); 1427 _STLP_TRY { 1428 __remainder = _S_new_RopeLeaf(__rest_buffer, __rest, __a); 1429 } 1430 _STLP_UNWIND(_RopeRep::_S_free_string(__rest_buffer, __rest, __a)) 1431 } 1432 __remainder_rope._M_tree_ptr._M_data = __remainder; 1433 if (__exponent != 0) { 1434 _CharT* __base_buffer = _M_tree_ptr.allocate(_S_rounded_up_size(__exponentiate_threshold)); 1435 _RopeLeaf* __base_leaf; 1436 rope<_CharT,_Alloc> __base_rope; 1437 uninitialized_fill_n(__base_buffer, __exponentiate_threshold, __c); 1438 _S_construct_null(__base_buffer + __exponentiate_threshold); 1439 _STLP_TRY { 1440 __base_leaf = _S_new_RopeLeaf(__base_buffer, 1441 __exponentiate_threshold, __a); 1442 } 1443 _STLP_UNWIND(_RopeRep::_S_free_string(__base_buffer, 1444 __exponentiate_threshold, __a)) 1445 __base_rope._M_tree_ptr._M_data = __base_leaf; 1446 if (1 == __exponent) { 1447 __result = __base_rope; 1448 // One each for base_rope and __result 1449 //_STLP_ASSERT(2 == __result._M_tree_ptr._M_data->_M_ref_count) 1450 } else { 1451 __result = _STLP_PRIV __power(__base_rope, __exponent, _Concat_fn()); 1452 } 1453 if (0 != __remainder) { 1454 __result += __remainder_rope; 1455 } 1456 } else { 1457 __result = __remainder_rope; 1458 } 1459 _M_tree_ptr._M_data = __result._M_tree_ptr._M_data; 1460 _M_tree_ptr._M_data->_M_ref_nonnil(); 1461 # undef __exponentiate_threshold 1462 } 1463 1464 rope(const allocator_type& __a = allocator_type()) 1465 : _M_tree_ptr(__a, (_RopeRep*)0) {} 1466 1467 // Construct a rope from a function that can compute its members 1468 rope(char_producer<_CharT> *__fn, size_t __len, bool __delete_fn, 1469 const allocator_type& __a = allocator_type()) 1470 : _M_tree_ptr(__a, (_RopeRep*)0) { 1471 _M_tree_ptr._M_data = (0 == __len) ? 1472 0 : _S_new_RopeFunction(__fn, __len, __delete_fn, __a); 1473 } 1474 1475 rope(const _Self& __x) 1476 : _M_tree_ptr(__x._M_tree_ptr, __x._M_tree_ptr._M_data) { 1477 _S_ref(_M_tree_ptr._M_data); 1478 } 1479 1480 #if !defined (_STLP_NO_MOVE_SEMANTIC) 1481 rope(__move_source<_Self> __src) 1482 : _M_tree_ptr(__src.get()._M_tree_ptr, __src.get()._M_tree_ptr._M_data) { 1483 __src.get()._M_tree_ptr._M_data = 0; 1484 } 1485 #endif 1486 1487 ~rope() { 1488 _S_unref(_M_tree_ptr._M_data); 1489 } 1490 1491 _Self& operator=(const _Self& __x) { 1492 _STLP_ASSERT(get_allocator() == __x.get_allocator()) 1493 _S_ref(__x._M_tree_ptr._M_data); 1494 _M_reset(__x._M_tree_ptr._M_data); 1495 return *this; 1496 } 1497 1498 void clear() { 1499 _S_unref(_M_tree_ptr._M_data); 1500 _M_tree_ptr._M_data = 0; 1501 } 1502 void push_back(_CharT __x) { 1503 _M_reset(_S_destr_concat_char_iter(_M_tree_ptr._M_data, &__x, 1)); 1504 } 1505 1506 void pop_back() { 1507 _RopeRep* __old = _M_tree_ptr._M_data; 1508 _M_tree_ptr._M_data = 1509 _S_substring(_M_tree_ptr._M_data, 0, _M_tree_ptr._M_data->_M_size._M_data - 1); 1510 _S_unref(__old); 1511 } 1512 1513 _CharT back() const { 1514 return _S_fetch(_M_tree_ptr._M_data, _M_tree_ptr._M_data->_M_size._M_data - 1); 1515 } 1516 1517 void push_front(_CharT __x) { 1518 _RopeRep* __old = _M_tree_ptr._M_data; 1519 _RopeRep* __left = 1520 _S_RopeLeaf_from_unowned_char_ptr(&__x, 1, _M_tree_ptr); 1521 _STLP_TRY { 1522 _M_tree_ptr._M_data = _S_concat_rep(__left, _M_tree_ptr._M_data); 1523 _S_unref(__old); 1524 _S_unref(__left); 1525 } 1526 _STLP_UNWIND(_S_unref(__left)) 1527 } 1528 1529 void pop_front() { 1530 _RopeRep* __old = _M_tree_ptr._M_data; 1531 _M_tree_ptr._M_data = _S_substring(_M_tree_ptr._M_data, 1, _M_tree_ptr._M_data->_M_size._M_data); 1532 _S_unref(__old); 1533 } 1534 1535 _CharT front() const { 1536 return _S_fetch(_M_tree_ptr._M_data, 0); 1537 } 1538 1539 void balance() { 1540 _RopeRep* __old = _M_tree_ptr._M_data; 1541 _M_tree_ptr._M_data = _S_balance(_M_tree_ptr._M_data); 1542 _S_unref(__old); 1543 } 1544 1545 void copy(_CharT* __buffer) const { 1546 _STLP_STD::_Destroy_Range(__buffer, __buffer + size()); 1547 _S_flatten(_M_tree_ptr._M_data, __buffer); 1548 } 1549 1550 /* 1551 * This is the copy function from the standard, but 1552 * with the arguments reordered to make it consistent with the 1553 * rest of the interface. 1554 * Note that this guaranteed not to compile if the draft standard 1555 * order is assumed. 1556 */ 1557 size_type copy(size_type __pos, size_type __n, _CharT* __buffer) const { 1558 size_t _p_size = size(); 1559 size_t __len = (__pos + __n > _p_size? _p_size - __pos : __n); 1560 1561 _STLP_STD::_Destroy_Range(__buffer, __buffer + __len); 1562 _S_flatten(_M_tree_ptr._M_data, __pos, __len, __buffer); 1563 return __len; 1564 } 1565 1566 # ifdef _STLP_DEBUG 1567 // Print to stdout, exposing structure. May be useful for 1568 // performance debugging. 1569 void dump() { 1570 _S_dump(_M_tree_ptr._M_data); 1571 } 1572 # endif 1573 1574 // Convert to 0 terminated string in new allocated memory. 1575 // Embedded 0s in the input do not terminate the copy. 1576 const _CharT* c_str() const; 1577 1578 // As above, but also use the flattened representation as the 1579 // the new rope representation. 1580 const _CharT* replace_with_c_str(); 1581 1582 // Reclaim memory for the c_str generated flattened string. 1583 // Intentionally undocumented, since it's hard to say when this 1584 // is safe for multiple threads. 1585 void delete_c_str () { 1586 if (0 == _M_tree_ptr._M_data) return; 1587 if (_RopeRep::_S_leaf == _M_tree_ptr._M_data->_M_tag && 1588 ((_RopeLeaf*)_M_tree_ptr._M_data)->_M_data == 1589 _M_tree_ptr._M_data->_M_c_string) { 1590 // Representation shared 1591 return; 1592 } 1593 _M_tree_ptr._M_data->_M_free_c_string(); 1594 _M_tree_ptr._M_data->_M_c_string = 0; 1595 } 1596 1597 _CharT operator[] (size_type __pos) const { 1598 return _S_fetch(_M_tree_ptr._M_data, __pos); 1599 } 1600 1601 _CharT at(size_type __pos) const { 1602 if (__pos >= size()) _M_throw_out_of_range(); 1603 return (*this)[__pos]; 1604 } 1605 1606 const_iterator begin() const { 1607 return(const_iterator(_M_tree_ptr._M_data, 0)); 1608 } 1609 1610 // An easy way to get a const iterator from a non-const container. 1611 const_iterator const_begin() const { 1612 return(const_iterator(_M_tree_ptr._M_data, 0)); 1613 } 1614 1615 const_iterator end() const { 1616 return(const_iterator(_M_tree_ptr._M_data, size())); 1617 } 1618 1619 const_iterator const_end() const { 1620 return(const_iterator(_M_tree_ptr._M_data, size())); 1621 } 1622 1623 size_type size() const { 1624 return(0 == _M_tree_ptr._M_data? 0 : _M_tree_ptr._M_data->_M_size._M_data); 1625 } 1626 1627 size_type length() const { 1628 return size(); 1629 } 1630 1631 size_type max_size() const { 1632 return _S_min_len[__ROPE_MAX_DEPTH-1] - 1; 1633 // Guarantees that the result can be sufficiently 1634 // balanced. Longer ropes will probably still work, 1635 // but it's harder to make guarantees. 1636 } 1637 1638 const_reverse_iterator rbegin() const { 1639 return const_reverse_iterator(end()); 1640 } 1641 1642 const_reverse_iterator const_rbegin() const { 1643 return const_reverse_iterator(end()); 1644 } 1645 1646 const_reverse_iterator rend() const { 1647 return const_reverse_iterator(begin()); 1648 } 1649 1650 const_reverse_iterator const_rend() const { 1651 return const_reverse_iterator(begin()); 1652 } 1653 // The symmetric cases are intentionally omitted, since they're presumed 1654 // to be less common, and we don't handle them as well. 1655 1656 // The following should really be templatized. 1657 // The first argument should be an input iterator or 1658 // forward iterator with value_type _CharT. 1659 _Self& append(const _CharT* __iter, size_t __n) { 1660 _M_reset(_S_destr_concat_char_iter(_M_tree_ptr._M_data, __iter, __n)); 1661 return *this; 1662 } 1663 1664 _Self& append(const _CharT* __c_string) { 1665 size_t __len = _S_char_ptr_len(__c_string); 1666 append(__c_string, __len); 1667 return *this; 1668 } 1669 1670 _Self& append(const _CharT* __s, const _CharT* __e) { 1671 _M_reset(_S_destr_concat_char_iter(_M_tree_ptr._M_data, __s, __e - __s)); 1672 return *this; 1673 } 1674 1675 _Self& append(const_iterator __s, const_iterator __e) { 1676 _STLP_ASSERT(__s._M_root == __e._M_root) 1677 _STLP_ASSERT(get_allocator() == __s._M_root->get_allocator()) 1678 _Self_destruct_ptr __appendee(_S_substring(__s._M_root, __s._M_current_pos, __e._M_current_pos)); 1679 _M_reset(_S_concat_rep(_M_tree_ptr._M_data, (_RopeRep*)__appendee)); 1680 return *this; 1681 } 1682 1683 _Self& append(_CharT __c) { 1684 _M_reset(_S_destr_concat_char_iter(_M_tree_ptr._M_data, &__c, 1)); 1685 return *this; 1686 } 1687 1688 _Self& append() { return append(_CharT()); } // XXX why? 1689 1690 _Self& append(const _Self& __y) { 1691 _STLP_ASSERT(__y.get_allocator() == get_allocator()) 1692 _M_reset(_S_concat_rep(_M_tree_ptr._M_data, __y._M_tree_ptr._M_data)); 1693 return *this; 1694 } 1695 1696 _Self& append(size_t __n, _CharT __c) { 1697 rope<_CharT,_Alloc> __last(__n, __c); 1698 return append(__last); 1699 } 1700 1701 void swap(_Self& __b) { 1702 _M_tree_ptr.swap(__b._M_tree_ptr); 1703 } 1704 #if defined (_STLP_USE_PARTIAL_SPEC_WORKAROUND) && !defined (_STLP_FUNCTION_TMPL_PARTIAL_ORDER) 1705 void _M_swap_workaround(_Self& __x) { swap(__x); } 1706 #endif 1707 1708 protected: 1709 // Result is included in refcount. 1710 static _RopeRep* replace(_RopeRep* __old, size_t __pos1, 1711 size_t __pos2, _RopeRep* __r) { 1712 if (0 == __old) { _S_ref(__r); return __r; } 1713 _Self_destruct_ptr __left(_S_substring(__old, 0, __pos1)); 1714 _Self_destruct_ptr __right(_S_substring(__old, __pos2, __old->_M_size._M_data)); 1715 _STLP_MPWFIX_TRY //*TY 06/01/2000 - 1716 _RopeRep* __result; 1717 1718 if (0 == __r) { 1719 __result = _S_concat_rep(__left, __right); 1720 } else { 1721 _STLP_ASSERT(__old->get_allocator() == __r->get_allocator()) 1722 _Self_destruct_ptr __left_result(_S_concat_rep(__left, __r)); 1723 __result = _S_concat_rep(__left_result, __right); 1724 } 1725 return __result; 1726 _STLP_MPWFIX_CATCH //*TY 06/01/2000 - 1727 } 1728 1729 public: 1730 void insert(size_t __p, const _Self& __r) { 1731 if (__p > size()) _M_throw_out_of_range(); 1732 _STLP_ASSERT(get_allocator() == __r.get_allocator()) 1733 _M_reset(replace(_M_tree_ptr._M_data, __p, __p, __r._M_tree_ptr._M_data)); 1734 } 1735 1736 void insert(size_t __p, size_t __n, _CharT __c) { 1737 rope<_CharT,_Alloc> __r(__n,__c); 1738 insert(__p, __r); 1739 } 1740 1741 void insert(size_t __p, const _CharT* __i, size_t __n) { 1742 if (__p > size()) _M_throw_out_of_range(); 1743 _Self_destruct_ptr __left(_S_substring(_M_tree_ptr._M_data, 0, __p)); 1744 _Self_destruct_ptr __right(_S_substring(_M_tree_ptr._M_data, __p, size())); 1745 _Self_destruct_ptr __left_result( 1746 _S_concat_char_iter(__left, __i, __n)); 1747 // _S_ destr_concat_char_iter should be safe here. 1748 // But as it stands it's probably not a win, since __left 1749 // is likely to have additional references. 1750 _M_reset(_S_concat_rep(__left_result, __right)); 1751 } 1752 1753 void insert(size_t __p, const _CharT* __c_string) { 1754 insert(__p, __c_string, _S_char_ptr_len(__c_string)); 1755 } 1756 1757 void insert(size_t __p, _CharT __c) { 1758 insert(__p, &__c, 1); 1759 } 1760 1761 void insert(size_t __p) { 1762 _CharT __c = _CharT(); 1763 insert(__p, &__c, 1); 1764 } 1765 1766 void insert(size_t __p, const _CharT* __i, const _CharT* __j) { 1767 _Self __r(__i, __j); 1768 insert(__p, __r); 1769 } 1770 1771 void insert(size_t __p, const const_iterator& __i, 1772 const const_iterator& __j) { 1773 _Self __r(__i, __j); 1774 insert(__p, __r); 1775 } 1776 1777 void insert(size_t __p, const iterator& __i, 1778 const iterator& __j) { 1779 _Self __r(__i, __j); 1780 insert(__p, __r); 1781 } 1782 1783 // (position, length) versions of replace operations: 1784 void replace(size_t __p, size_t __n, const _Self& __r) { 1785 if (__p > size()) _M_throw_out_of_range(); 1786 _M_reset(replace(_M_tree_ptr._M_data, __p, __p + __n, __r._M_tree_ptr._M_data)); 1787 } 1788 1789 void replace(size_t __p, size_t __n, 1790 const _CharT* __i, size_t __i_len) { 1791 _Self __r(__i, __i_len); 1792 replace(__p, __n, __r); 1793 } 1794 1795 void replace(size_t __p, size_t __n, _CharT __c) { 1796 _Self __r(__c); 1797 replace(__p, __n, __r); 1798 } 1799 1800 void replace(size_t __p, size_t __n, const _CharT* __c_string) { 1801 _Self __r(__c_string); 1802 replace(__p, __n, __r); 1803 } 1804 1805 void replace(size_t __p, size_t __n, 1806 const _CharT* __i, const _CharT* __j) { 1807 _Self __r(__i, __j); 1808 replace(__p, __n, __r); 1809 } 1810 1811 void replace(size_t __p, size_t __n, 1812 const const_iterator& __i, const const_iterator& __j) { 1813 _Self __r(__i, __j); 1814 replace(__p, __n, __r); 1815 } 1816 1817 void replace(size_t __p, size_t __n, 1818 const iterator& __i, const iterator& __j) { 1819 _Self __r(__i, __j); 1820 replace(__p, __n, __r); 1821 } 1822 1823 // Single character variants: 1824 void replace(size_t __p, _CharT __c) { 1825 if (__p > size()) _M_throw_out_of_range(); 1826 iterator __i(this, __p); 1827 *__i = __c; 1828 } 1829 1830 void replace(size_t __p, const _Self& __r) { 1831 replace(__p, 1, __r); 1832 } 1833 1834 void replace(size_t __p, const _CharT* __i, size_t __i_len) { 1835 replace(__p, 1, __i, __i_len); 1836 } 1837 1838 void replace(size_t __p, const _CharT* __c_string) { 1839 replace(__p, 1, __c_string); 1840 } 1841 1842 void replace(size_t __p, const _CharT* __i, const _CharT* __j) { 1843 replace(__p, 1, __i, __j); 1844 } 1845 1846 void replace(size_t __p, const const_iterator& __i, 1847 const const_iterator& __j) { 1848 replace(__p, 1, __i, __j); 1849 } 1850 1851 void replace(size_t __p, const iterator& __i, 1852 const iterator& __j) { 1853 replace(__p, 1, __i, __j); 1854 } 1855 1856 // Erase, (position, size) variant. 1857 void erase(size_t __p, size_t __n) { 1858 if (__p > size()) _M_throw_out_of_range(); 1859 _M_reset(replace(_M_tree_ptr._M_data, __p, __p + __n, 0)); 1860 } 1861 1862 // Erase, single character 1863 void erase(size_t __p) { 1864 erase(__p, __p + 1); 1865 } 1866 1867 // Insert, iterator variants. 1868 iterator insert(const iterator& __p, const _Self& __r) 1869 { insert(__p.index(), __r); return __p; } 1870 iterator insert(const iterator& __p, size_t __n, _CharT __c) 1871 { insert(__p.index(), __n, __c); return __p; } 1872 iterator insert(const iterator& __p, _CharT __c) 1873 { insert(__p.index(), __c); return __p; } 1874 iterator insert(const iterator& __p ) 1875 { insert(__p.index()); return __p; } 1876 iterator insert(const iterator& __p, const _CharT* c_string) 1877 { insert(__p.index(), c_string); return __p; } 1878 iterator insert(const iterator& __p, const _CharT* __i, size_t __n) 1879 { insert(__p.index(), __i, __n); return __p; } 1880 iterator insert(const iterator& __p, const _CharT* __i, 1881 const _CharT* __j) 1882 { insert(__p.index(), __i, __j); return __p; } 1883 iterator insert(const iterator& __p, 1884 const const_iterator& __i, const const_iterator& __j) 1885 { insert(__p.index(), __i, __j); return __p; } 1886 iterator insert(const iterator& __p, 1887 const iterator& __i, const iterator& __j) 1888 { insert(__p.index(), __i, __j); return __p; } 1889 1890 // Replace, range variants. 1891 void replace(const iterator& __p, const iterator& __q, 1892 const _Self& __r) 1893 { replace(__p.index(), __q.index() - __p.index(), __r); } 1894 void replace(const iterator& __p, const iterator& __q, _CharT __c) 1895 { replace(__p.index(), __q.index() - __p.index(), __c); } 1896 void replace(const iterator& __p, const iterator& __q, 1897 const _CharT* __c_string) 1898 { replace(__p.index(), __q.index() - __p.index(), __c_string); } 1899 void replace(const iterator& __p, const iterator& __q, 1900 const _CharT* __i, size_t __n) 1901 { replace(__p.index(), __q.index() - __p.index(), __i, __n); } 1902 void replace(const iterator& __p, const iterator& __q, 1903 const _CharT* __i, const _CharT* __j) 1904 { replace(__p.index(), __q.index() - __p.index(), __i, __j); } 1905 void replace(const iterator& __p, const iterator& __q, 1906 const const_iterator& __i, const const_iterator& __j) 1907 { replace(__p.index(), __q.index() - __p.index(), __i, __j); } 1908 void replace(const iterator& __p, const iterator& __q, 1909 const iterator& __i, const iterator& __j) 1910 { replace(__p.index(), __q.index() - __p.index(), __i, __j); } 1911 1912 // Replace, iterator variants. 1913 void replace(const iterator& __p, const _Self& __r) 1914 { replace(__p.index(), __r); } 1915 void replace(const iterator& __p, _CharT __c) 1916 { replace(__p.index(), __c); } 1917 void replace(const iterator& __p, const _CharT* __c_string) 1918 { replace(__p.index(), __c_string); } 1919 void replace(const iterator& __p, const _CharT* __i, size_t __n) 1920 { replace(__p.index(), __i, __n); } 1921 void replace(const iterator& __p, const _CharT* __i, const _CharT* __j) 1922 { replace(__p.index(), __i, __j); } 1923 void replace(const iterator& __p, const_iterator __i, 1924 const_iterator __j) 1925 { replace(__p.index(), __i, __j); } 1926 void replace(const iterator& __p, iterator __i, iterator __j) 1927 { replace(__p.index(), __i, __j); } 1928 1929 // Iterator and range variants of erase 1930 iterator erase(const iterator& __p, const iterator& __q) { 1931 size_t __p_index = __p.index(); 1932 erase(__p_index, __q.index() - __p_index); 1933 return iterator(this, __p_index); 1934 } 1935 iterator erase(const iterator& __p) { 1936 size_t __p_index = __p.index(); 1937 erase(__p_index, 1); 1938 return iterator(this, __p_index); 1939 } 1940 1941 _Self substr(size_t __start, size_t __len = 1) const { 1942 if (__start > size()) _M_throw_out_of_range(); 1943 return rope<_CharT,_Alloc>(_S_substring(_M_tree_ptr._M_data, __start, __start + __len)); 1944 } 1945 1946 _Self substr(iterator __start, iterator __end) const { 1947 return rope<_CharT,_Alloc>(_S_substring(_M_tree_ptr._M_data, __start.index(), __end.index())); 1948 } 1949 1950 _Self substr(iterator __start) const { 1951 size_t __pos = __start.index(); 1952 return rope<_CharT,_Alloc>(_S_substring(_M_tree_ptr._M_data, __pos, __pos + 1)); 1953 } 1954 1955 _Self substr(const_iterator __start, const_iterator __end) const { 1956 // This might eventually take advantage of the cache in the 1957 // iterator. 1958 return rope<_CharT,_Alloc>(_S_substring(_M_tree_ptr._M_data, __start.index(), __end.index())); 1959 } 1960 1961 rope<_CharT,_Alloc> substr(const_iterator __start) { 1962 size_t __pos = __start.index(); 1963 return rope<_CharT,_Alloc>(_S_substring(_M_tree_ptr._M_data, __pos, __pos + 1)); 1964 } 1965 1966 #include <stl/_string_npos.h> 1967 1968 size_type find(const _Self& __s, size_type __pos = 0) const { 1969 if (__pos >= size()) 1970 # ifndef _STLP_OLD_ROPE_SEMANTICS 1971 return npos; 1972 # else 1973 return size(); 1974 # endif 1975 1976 size_type __result_pos; 1977 const_iterator __result = _STLP_STD::search(const_begin() + (ptrdiff_t)__pos, const_end(), __s.begin(), __s.end() ); 1978 __result_pos = __result.index(); 1979 # ifndef _STLP_OLD_ROPE_SEMANTICS 1980 if (__result_pos == size()) __result_pos = npos; 1981 # endif 1982 return __result_pos; 1983 } 1984 size_type find(_CharT __c, size_type __pos = 0) const; 1985 size_type find(const _CharT* __s, size_type __pos = 0) const { 1986 size_type __result_pos; 1987 const_iterator __result = _STLP_STD::search(const_begin() + (ptrdiff_t)__pos, const_end(), 1988 __s, __s + _S_char_ptr_len(__s)); 1989 __result_pos = __result.index(); 1990 # ifndef _STLP_OLD_ROPE_SEMANTICS 1991 if (__result_pos == size()) __result_pos = npos; 1992 # endif 1993 return __result_pos; 1994 } 1995 1996 iterator mutable_begin() { 1997 return(iterator(this, 0)); 1998 } 1999 2000 iterator mutable_end() { 2001 return(iterator(this, size())); 2002 } 2003 2004 reverse_iterator mutable_rbegin() { 2005 return reverse_iterator(mutable_end()); 2006 } 2007 2008 reverse_iterator mutable_rend() { 2009 return reverse_iterator(mutable_begin()); 2010 } 2011 2012 reference mutable_reference_at(size_type __pos) { 2013 return reference(this, __pos); 2014 } 2015 2016 # ifdef __STD_STUFF 2017 reference operator[] (size_type __pos) { 2018 return reference(this, __pos); 2019 } 2020 2021 reference at(size_type __pos) { 2022 if (__pos >= size()) _M_throw_out_of_range(); 2023 return (*this)[__pos]; 2024 } 2025 2026 void resize(size_type, _CharT) {} 2027 void resize(size_type) {} 2028 void reserve(size_type = 0) {} 2029 size_type capacity() const { 2030 return max_size(); 2031 } 2032 2033 // Stuff below this line is dangerous because it's error prone. 2034 // I would really like to get rid of it. 2035 // copy function with funny arg ordering. 2036 size_type copy(_CharT* __buffer, size_type __n, 2037 size_type __pos = 0) const { 2038 return copy(__pos, __n, __buffer); 2039 } 2040 2041 iterator end() { return mutable_end(); } 2042 2043 iterator begin() { return mutable_begin(); } 2044 2045 reverse_iterator rend() { return mutable_rend(); } 2046 2047 reverse_iterator rbegin() { return mutable_rbegin(); } 2048 2049 # else 2050 2051 const_iterator end() { return const_end(); } 2052 2053 const_iterator begin() { return const_begin(); } 2054 2055 const_reverse_iterator rend() { return const_rend(); } 2056 2057 const_reverse_iterator rbegin() { return const_rbegin(); } 2058 2059 # endif 2060 }; //class rope 2061 2062 #if defined (__GNUC__) && (__GNUC__ == 2) && (__GNUC_MINOR__ == 96) 2063 template <class _CharT, class _Alloc> 2064 const size_t rope<_CharT, _Alloc>::npos = ~(size_t) 0; 2065 #endif 2066 2067 template <class _CharT, class _Alloc> 2068 inline _CharT 2069 _Rope_const_iterator< _CharT, _Alloc>::operator[](size_t __n) 2070 { return rope<_CharT,_Alloc>::_S_fetch(this->_M_root, this->_M_current_pos + __n); } 2071 2072 template <class _CharT, class _Alloc> 2073 inline bool operator== (const _Rope_const_iterator<_CharT,_Alloc>& __x, 2074 const _Rope_const_iterator<_CharT,_Alloc>& __y) { 2075 return (__x._M_current_pos == __y._M_current_pos && 2076 __x._M_root == __y._M_root); 2077 } 2078 2079 template <class _CharT, class _Alloc> 2080 inline bool operator< (const _Rope_const_iterator<_CharT,_Alloc>& __x, 2081 const _Rope_const_iterator<_CharT,_Alloc>& __y) 2082 { return (__x._M_current_pos < __y._M_current_pos); } 2083 2084 #ifdef _STLP_USE_SEPARATE_RELOPS_NAMESPACE 2085 2086 template <class _CharT, class _Alloc> 2087 inline bool operator!= (const _Rope_const_iterator<_CharT,_Alloc>& __x, 2088 const _Rope_const_iterator<_CharT,_Alloc>& __y) 2089 { return !(__x == __y); } 2090 2091 template <class _CharT, class _Alloc> 2092 inline bool operator> (const _Rope_const_iterator<_CharT,_Alloc>& __x, 2093 const _Rope_const_iterator<_CharT,_Alloc>& __y) 2094 { return __y < __x; } 2095 2096 template <class _CharT, class _Alloc> 2097 inline bool operator<= (const _Rope_const_iterator<_CharT,_Alloc>& __x, 2098 const _Rope_const_iterator<_CharT,_Alloc>& __y) 2099 { return !(__y < __x); } 2100 2101 template <class _CharT, class _Alloc> 2102 inline bool operator>= (const _Rope_const_iterator<_CharT,_Alloc>& __x, 2103 const _Rope_const_iterator<_CharT,_Alloc>& __y) 2104 { return !(__x < __y); } 2105 2106 #endif /* _STLP_USE_SEPARATE_RELOPS_NAMESPACE */ 2107 2108 template <class _CharT, class _Alloc> 2109 inline ptrdiff_t operator-(const _Rope_const_iterator<_CharT,_Alloc>& __x, 2110 const _Rope_const_iterator<_CharT,_Alloc>& __y) 2111 { return (ptrdiff_t)__x._M_current_pos - (ptrdiff_t)__y._M_current_pos; } 2112 2113 #if !defined( __MWERKS__ ) || __MWERKS__ >= 0x2000 // dwa 8/21/97 - "ambiguous access to overloaded function" bug. 2114 template <class _CharT, class _Alloc> 2115 inline _Rope_const_iterator<_CharT,_Alloc> 2116 operator-(const _Rope_const_iterator<_CharT,_Alloc>& __x, ptrdiff_t __n) 2117 { return _Rope_const_iterator<_CharT,_Alloc>(__x._M_root, __x._M_current_pos - __n); } 2118 # endif 2119 2120 template <class _CharT, class _Alloc> 2121 inline _Rope_const_iterator<_CharT,_Alloc> 2122 operator+(const _Rope_const_iterator<_CharT,_Alloc>& __x, ptrdiff_t __n) 2123 { return _Rope_const_iterator<_CharT,_Alloc>(__x._M_root, __x._M_current_pos + __n); } 2124 2125 template <class _CharT, class _Alloc> 2126 inline _Rope_const_iterator<_CharT,_Alloc> 2127 operator+(ptrdiff_t __n, const _Rope_const_iterator<_CharT,_Alloc>& __x) 2128 { return _Rope_const_iterator<_CharT,_Alloc>(__x._M_root, __x._M_current_pos + __n); } 2129 2130 template <class _CharT, class _Alloc> 2131 inline bool operator== (const _Rope_iterator<_CharT,_Alloc>& __x, 2132 const _Rope_iterator<_CharT,_Alloc>& __y) { 2133 return (__x._M_current_pos == __y._M_current_pos && 2134 __x._M_root_rope == __y._M_root_rope); 2135 } 2136 2137 template <class _CharT, class _Alloc> 2138 inline bool operator< (const _Rope_iterator<_CharT,_Alloc>& __x, 2139 const _Rope_iterator<_CharT,_Alloc>& __y) 2140 { return (__x._M_current_pos < __y._M_current_pos); } 2141 2142 #if defined (_STLP_USE_SEPARATE_RELOPS_NAMESPACE) 2143 template <class _CharT, class _Alloc> 2144 inline bool operator!= (const _Rope_iterator<_CharT,_Alloc>& __x, 2145 const _Rope_iterator<_CharT,_Alloc>& __y) 2146 { return !(__x == __y); } 2147 2148 template <class _CharT, class _Alloc> 2149 inline bool operator> (const _Rope_iterator<_CharT,_Alloc>& __x, 2150 const _Rope_iterator<_CharT,_Alloc>& __y) 2151 { return __y < __x; } 2152 2153 template <class _CharT, class _Alloc> 2154 inline bool operator<= (const _Rope_iterator<_CharT,_Alloc>& __x, 2155 const _Rope_iterator<_CharT,_Alloc>& __y) 2156 { return !(__y < __x); } 2157 2158 template <class _CharT, class _Alloc> 2159 inline bool operator>= (const _Rope_iterator<_CharT,_Alloc>& __x, 2160 const _Rope_iterator<_CharT,_Alloc>& __y) 2161 { return !(__x < __y); } 2162 #endif /* _STLP_USE_SEPARATE_RELOPS_NAMESPACE */ 2163 2164 template <class _CharT, class _Alloc> 2165 inline ptrdiff_t operator-(const _Rope_iterator<_CharT,_Alloc>& __x, 2166 const _Rope_iterator<_CharT,_Alloc>& __y) 2167 { return (ptrdiff_t)__x._M_current_pos - (ptrdiff_t)__y._M_current_pos; } 2168 2169 #if !defined( __MWERKS__ ) || __MWERKS__ >= 0x2000 // dwa 8/21/97 - "ambiguous access to overloaded function" bug. 2170 template <class _CharT, class _Alloc> 2171 inline _Rope_iterator<_CharT,_Alloc> 2172 operator-(const _Rope_iterator<_CharT,_Alloc>& __x, 2173 ptrdiff_t __n) { 2174 return _Rope_iterator<_CharT,_Alloc>(__x._M_root_rope, __x._M_current_pos - __n); 2175 } 2176 # endif 2177 2178 template <class _CharT, class _Alloc> 2179 inline _Rope_iterator<_CharT,_Alloc> 2180 operator+(const _Rope_iterator<_CharT,_Alloc>& __x, 2181 ptrdiff_t __n) { 2182 return _Rope_iterator<_CharT,_Alloc>(__x._M_root_rope, __x._M_current_pos + __n); 2183 } 2184 2185 template <class _CharT, class _Alloc> 2186 inline _Rope_iterator<_CharT,_Alloc> 2187 operator+(ptrdiff_t __n, const _Rope_iterator<_CharT,_Alloc>& __x) { 2188 return _Rope_iterator<_CharT,_Alloc>(__x._M_root_rope, __x._M_current_pos + __n); 2189 } 2190 2191 template <class _CharT, class _Alloc> 2192 inline rope<_CharT,_Alloc> 2193 operator+ (const rope<_CharT,_Alloc>& __left, 2194 const rope<_CharT,_Alloc>& __right) { 2195 _STLP_ASSERT(__left.get_allocator() == __right.get_allocator()) 2196 return rope<_CharT,_Alloc>(rope<_CharT,_Alloc>::_S_concat_rep(__left._M_tree_ptr._M_data, __right._M_tree_ptr._M_data)); 2197 // Inlining this should make it possible to keep __left and __right in registers. 2198 } 2199 2200 template <class _CharT, class _Alloc> 2201 inline rope<_CharT,_Alloc>& 2202 operator+= (rope<_CharT,_Alloc>& __left, 2203 const rope<_CharT,_Alloc>& __right) { 2204 __left.append(__right); 2205 return __left; 2206 } 2207 2208 template <class _CharT, class _Alloc> 2209 inline rope<_CharT,_Alloc> 2210 operator+ (const rope<_CharT,_Alloc>& __left, 2211 const _CharT* __right) { 2212 size_t __rlen = rope<_CharT,_Alloc>::_S_char_ptr_len(__right); 2213 return rope<_CharT,_Alloc>(rope<_CharT,_Alloc>::_S_concat_char_iter(__left._M_tree_ptr._M_data, __right, __rlen)); 2214 } 2215 2216 template <class _CharT, class _Alloc> 2217 inline rope<_CharT,_Alloc>& 2218 operator+= (rope<_CharT,_Alloc>& __left, 2219 const _CharT* __right) { 2220 __left.append(__right); 2221 return __left; 2222 } 2223 2224 template <class _CharT, class _Alloc> 2225 inline rope<_CharT,_Alloc> 2226 operator+ (const rope<_CharT,_Alloc>& __left, _CharT __right) { 2227 return rope<_CharT,_Alloc>(rope<_CharT,_Alloc>::_S_concat_char_iter(__left._M_tree_ptr._M_data, &__right, 1)); 2228 } 2229 2230 template <class _CharT, class _Alloc> 2231 inline rope<_CharT,_Alloc>& 2232 operator+= (rope<_CharT,_Alloc>& __left, _CharT __right) { 2233 __left.append(__right); 2234 return __left; 2235 } 2236 2237 template <class _CharT, class _Alloc> 2238 inline bool 2239 operator< (const rope<_CharT,_Alloc>& __left, 2240 const rope<_CharT,_Alloc>& __right) { 2241 return __left.compare(__right) < 0; 2242 } 2243 2244 template <class _CharT, class _Alloc> 2245 inline bool 2246 operator== (const rope<_CharT,_Alloc>& __left, 2247 const rope<_CharT,_Alloc>& __right) { 2248 return __left.compare(__right) == 0; 2249 } 2250 2251 #ifdef _STLP_USE_SEPARATE_RELOPS_NAMESPACE 2252 2253 template <class _CharT, class _Alloc> 2254 inline bool 2255 operator!= (const rope<_CharT,_Alloc>& __x, const rope<_CharT,_Alloc>& __y) { 2256 return !(__x == __y); 2257 } 2258 2259 template <class _CharT, class _Alloc> 2260 inline bool 2261 operator> (const rope<_CharT,_Alloc>& __x, const rope<_CharT,_Alloc>& __y) { 2262 return __y < __x; 2263 } 2264 2265 template <class _CharT, class _Alloc> 2266 inline bool 2267 operator<= (const rope<_CharT,_Alloc>& __x, const rope<_CharT,_Alloc>& __y) { 2268 return !(__y < __x); 2269 } 2270 2271 template <class _CharT, class _Alloc> 2272 inline bool 2273 operator>= (const rope<_CharT,_Alloc>& __x, const rope<_CharT,_Alloc>& __y) { 2274 return !(__x < __y); 2275 } 2276 2277 template <class _CharT, class _Alloc> 2278 inline bool operator!= (const _Rope_char_ptr_proxy<_CharT,_Alloc>& __x, 2279 const _Rope_char_ptr_proxy<_CharT,_Alloc>& __y) { 2280 return !(__x == __y); 2281 } 2282 2283 #endif /* _STLP_USE_SEPARATE_RELOPS_NAMESPACE */ 2284 2285 template <class _CharT, class _Alloc> 2286 inline bool operator== (const _Rope_char_ptr_proxy<_CharT,_Alloc>& __x, 2287 const _Rope_char_ptr_proxy<_CharT,_Alloc>& __y) { 2288 return (__x._M_pos == __y._M_pos && __x._M_root == __y._M_root); 2289 } 2290 2291 #if !defined (_STLP_USE_NO_IOSTREAMS) 2292 template<class _CharT, class _Traits, class _Alloc> 2293 basic_ostream<_CharT, _Traits>& operator<< (basic_ostream<_CharT, _Traits>& __o, 2294 const rope<_CharT, _Alloc>& __r); 2295 #endif 2296 2297 typedef rope<char, allocator<char> > crope; 2298 #if defined (_STLP_HAS_WCHAR_T) 2299 typedef rope<wchar_t, allocator<wchar_t> > wrope; 2300 #endif 2301 2302 inline crope::reference __mutable_reference_at(crope& __c, size_t __i) 2303 { return __c.mutable_reference_at(__i); } 2304 2305 #if defined (_STLP_HAS_WCHAR_T) 2306 inline wrope::reference __mutable_reference_at(wrope& __c, size_t __i) 2307 { return __c.mutable_reference_at(__i); } 2308 #endif 2309 2310 #if defined (_STLP_FUNCTION_TMPL_PARTIAL_ORDER) 2311 template <class _CharT, class _Alloc> 2312 inline void swap(rope<_CharT,_Alloc>& __x, rope<_CharT,_Alloc>& __y) 2313 { __x.swap(__y); } 2314 #else 2315 2316 inline void swap(crope& __x, crope& __y) { __x.swap(__y); } 2317 # ifdef _STLP_HAS_WCHAR_T // dwa 8/21/97 2318 inline void swap(wrope& __x, wrope& __y) { __x.swap(__y); } 2319 # endif 2320 2321 #endif /* _STLP_FUNCTION_TMPL_PARTIAL_ORDER */ 2322 2323 2324 // Hash functions should probably be revisited later: 2325 _STLP_TEMPLATE_NULL struct hash<crope> { 2326 size_t operator()(const crope& __str) const { 2327 size_t _p_size = __str.size(); 2328 2329 if (0 == _p_size) return 0; 2330 return 13*__str[0] + 5*__str[_p_size - 1] + _p_size; 2331 } 2332 }; 2333 2334 #if defined (_STLP_HAS_WCHAR_T) // dwa 8/21/97 2335 _STLP_TEMPLATE_NULL struct hash<wrope> { 2336 size_t operator()(const wrope& __str) const { 2337 size_t _p_size = __str.size(); 2338 2339 if (0 == _p_size) return 0; 2340 return 13*__str[0] + 5*__str[_p_size - 1] + _p_size; 2341 } 2342 }; 2343 #endif 2344 2345 #if (!defined (_STLP_MSVC) || (_STLP_MSVC >= 1310)) 2346 // I couldn't get this to work with VC++ 2347 template<class _CharT,class _Alloc> 2348 # if defined (__DMC__) 2349 extern 2350 # endif 2351 void _Rope_rotate(_Rope_iterator<_CharT, _Alloc> __first, 2352 _Rope_iterator<_CharT, _Alloc> __middle, 2353 _Rope_iterator<_CharT, _Alloc> __last); 2354 2355 inline void rotate(_Rope_iterator<char, allocator<char> > __first, 2356 _Rope_iterator<char, allocator<char> > __middle, 2357 _Rope_iterator<char, allocator<char> > __last) 2358 { _Rope_rotate(__first, __middle, __last); } 2359 #endif 2360 2361 template <class _CharT, class _Alloc> 2362 inline _Rope_char_ref_proxy<_CharT, _Alloc>::operator _CharT () const { 2363 if (_M_current_valid) { 2364 return _M_current; 2365 } else { 2366 return _My_rope::_S_fetch(_M_root->_M_tree_ptr._M_data, _M_pos); 2367 } 2368 } 2369 2370 #if defined (_STLP_CLASS_PARTIAL_SPECIALIZATION) && !defined (_STLP_NO_MOVE_SEMANTIC) 2371 template <class _CharT, class _Alloc> 2372 struct __move_traits<rope<_CharT, _Alloc> > { 2373 typedef __true_type implemented; 2374 //Completness depends on the allocator: 2375 typedef typename __move_traits<_Alloc>::complete complete; 2376 }; 2377 #endif 2378 2379 _STLP_END_NAMESPACE 2380 2381 #if !defined (_STLP_LINK_TIME_INSTANTIATION) 2382 # include <stl/_rope.c> 2383 #endif 2384 2385 #endif /* _STLP_INTERNAL_ROPE_H */ 2386 2387 // Local Variables: 2388 // mode:C++ 2389 // End: 2390