123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293 |
- // Copyright (c) 2001-2010 Hartmut Kaiser
- // Copyright (c) 2010 Mathias Gaunard
- //
- // Distributed under the Boost Software License, Version 1.0. (See accompanying
- // file LICENSE_1_0.txt or copy at http://www.boost.org/LICENSE_1_0.txt)
- #include <boost/config/warning_disable.hpp>
- #include <boost/detail/lightweight_test.hpp>
- #include <boost/spirit/include/support_multi_pass.hpp>
- #include <boost/spirit/include/classic_position_iterator.hpp>
- #include <boost/spirit/include/lex_lexertl.hpp>
- namespace spirit = boost::spirit;
- namespace lex = spirit::lex;
- typedef spirit::classic::position_iterator2<
- spirit::multi_pass<std::istreambuf_iterator<char> >
- > file_iterator;
- inline file_iterator
- make_file_iterator(std::istream& input, const std::string& filename)
- {
- return file_iterator(
- spirit::make_default_multi_pass(
- std::istreambuf_iterator<char>(input)),
- spirit::multi_pass<std::istreambuf_iterator<char> >(),
- filename);
- }
- enum token_id
- {
- ID_WORD = lex::min_token_id + 1,
- ID_EOL
- };
- typedef lex::lexertl::token<
- file_iterator, boost::mpl::vector<>, boost::mpl::true_, token_id
- > token_type;
- struct lexer
- : lex::lexer<lex::lexertl::actor_lexer<token_type> >
- {
- lexer() : word("^[a-zA-Z0-9]+$", ID_WORD)
- {
- typedef lex::token_def<lex::unused_type, char, token_id> toked_def;
- self("INITIAL", "O") =
- word
- | toked_def("!.*$") [
- lex::_pass = lex::pass_flags::pass_ignore
- ]
- | toked_def('\n', ID_EOL)
- ;
-
- self("O", "INITIAL") =
- toked_def(".") [
- lex::_pass = lex::pass_flags::pass_fail
- ]
- ;
- }
-
- lex::token_def<lex::unused_type, char, token_id> word;
- };
- typedef lexer::iterator_type token_iterator;
- int main()
- {
- std::stringstream ss;
- ss << "!foo\nbar\n!baz";
-
- file_iterator begin = make_file_iterator(ss, "SS");
- file_iterator end;
-
- lexer l;
- token_iterator begin2 = l.begin(begin, end);
- token_iterator end2 = l.end();
-
- token_id test_data[] = { ID_EOL, ID_WORD, ID_EOL };
- std::size_t const test_data_size = sizeof(test_data)/sizeof(test_data[0]);
- token_iterator it = begin2;
- std::size_t i = 0;
- for (/**/; it != end2 && i < test_data_size; ++it, ++i)
- {
- BOOST_TEST(it->id() == test_data[i]);
- }
- BOOST_TEST(it == end2);
- BOOST_TEST(i == test_data_size);
- return boost::report_errors();
- }
|