Re: [boost] Latest version of Lexertl

21 Nov
2007
21 Nov
'07
6:59 p.m.
Ben, > >> Yes, but as we said (unless you sent this before your other > >> reply) a 4.25MB (0x110000 * 4) lookup is definitely > unacceptable. I > >> suppose for the case of Windows (a 256K lookup - 65536 * > 4) this is > >> OK, so we could consider an auto compression scheme if you exceed > >> 65536 entries for the vector in question. Do you think > even 256K is > >> too much? > > > >2^16 (64k) is probably the max you should use to cut off if larger. > Remember > >the cache is not only there for the data, instructions are held in > there as > >well... > >Can you make this max size configurable? > > Sure, I'll need to understand how to do the compression first though. Sure. > >One last note: parser/tokeniser/str_token.hpp still includes > >size_t.h/consts.h (it should include size_t.hpp/consts.hpp), > same for > >examples/cpp_code.hpp and examples/csharp_code.hpp. > > My bad: just delete the directory tree (lexer) and > re-unzip... those files aren't used anymore. Ohh, that's too bad, because I'm using the cpp_code.hpp and serialise.hpp headers in the lexertl based lexer for Wave. Any chance to revive these? Regards Hartmut
6415
Age (days ago)
6415
Last active (days ago)
0 comments
1 participants
participants (1)
-
Hartmut Kaiser