aboutsummaryrefslogtreecommitdiff
path: root/paludis/repositories
diff options
context:
space:
mode:
Diffstat (limited to 'paludis/repositories')
-rw-r--r--paludis/repositories/cran/cran_dep_parser.cc2
-rw-r--r--paludis/repositories/cran/cran_package_id.cc2
-rw-r--r--paludis/repositories/e/dep_lexer.cc2
-rw-r--r--paludis/repositories/e/dep_parser.cc2
-rw-r--r--paludis/repositories/e/e_key.cc8
-rw-r--r--paludis/repositories/e/e_repository.cc9
-rw-r--r--paludis/repositories/e/e_repository_profile.cc20
-rw-r--r--paludis/repositories/e/eapi.cc10
-rw-r--r--paludis/repositories/e/eapi_phase.cc4
-rw-r--r--paludis/repositories/e/ebuild.cc2
-rw-r--r--paludis/repositories/e/ebuild_entries.cc2
-rw-r--r--paludis/repositories/e/ebuild_flat_metadata_cache.cc4
-rw-r--r--paludis/repositories/e/make_ebin_repository.cc2
-rw-r--r--paludis/repositories/e/make_ebuild_repository.cc4
-rw-r--r--paludis/repositories/e/manifest2_reader.cc3
-rw-r--r--paludis/repositories/e/qa/kv_variables.cc2
-rw-r--r--paludis/repositories/e/qa/visibility.cc2
-rw-r--r--paludis/repositories/e/qa/whitespace.cc2
-rw-r--r--paludis/repositories/e/vdb_merger.cc6
-rw-r--r--paludis/repositories/e/vdb_repository.cc2
-rw-r--r--paludis/repositories/e/vdb_unmerger.cc6
-rw-r--r--paludis/repositories/e/xml_things.cc5
-rw-r--r--paludis/repositories/fake/fake_package_id.cc4
-rw-r--r--paludis/repositories/unpackaged/dep_parser.cc2
-rw-r--r--paludis/repositories/unpackaged/ndbam.cc2
-rw-r--r--paludis/repositories/unpackaged/ndbam_merger.cc6
-rw-r--r--paludis/repositories/unpackaged/ndbam_unmerger.cc6
27 files changed, 55 insertions, 66 deletions
diff --git a/paludis/repositories/cran/cran_dep_parser.cc b/paludis/repositories/cran/cran_dep_parser.cc
index 3136c1a05..a37eeba88 100644
--- a/paludis/repositories/cran/cran_dep_parser.cc
+++ b/paludis/repositories/cran/cran_dep_parser.cc
@@ -39,7 +39,7 @@ cranrepository::parse_depends(const std::string & s)
new ConstTreeSequence<DependencySpecTree, AllDepSpec>(tr1::shared_ptr<AllDepSpec>(new AllDepSpec)));
std::list<std::string> specs;
- Tokeniser<delim_kind::AnyOfTag>::tokenise(s, ",", std::back_inserter(specs));
+ tokenise<delim_kind::AnyOfTag, delim_mode::DelimiterTag>(s, ",", "", std::back_inserter(specs));
std::list<std::string>::const_iterator a(specs.begin()), a_end(specs.end());
for ( ; a != a_end ; ++a)
diff --git a/paludis/repositories/cran/cran_package_id.cc b/paludis/repositories/cran/cran_package_id.cc
index 6443a4cca..07a9847e8 100644
--- a/paludis/repositories/cran/cran_package_id.cc
+++ b/paludis/repositories/cran/cran_package_id.cc
@@ -192,7 +192,7 @@ CRANPackageID::CRANPackageID(const Environment * const env, const tr1::shared_pt
{
Context local_context("When handling Contains: key:");
std::list<std::string> tokens;
- WhitespaceTokeniser::tokenise(file.get("Contains"), std::back_inserter(tokens));
+ tokenise_whitespace(file.get("Contains"), std::back_inserter(tokens));
_imp->contains_key.reset(new PackageIDSequenceKey(_imp->env, "Contains", "Contains", mkt_normal));
add_metadata_key(_imp->contains_key);
for (std::list<std::string>::const_iterator t(tokens.begin()), t_end(tokens.end()) ;
diff --git a/paludis/repositories/e/dep_lexer.cc b/paludis/repositories/e/dep_lexer.cc
index 90c6a7cf0..81fbe02b3 100644
--- a/paludis/repositories/e/dep_lexer.cc
+++ b/paludis/repositories/e/dep_lexer.cc
@@ -63,7 +63,7 @@ DepLexer::DepLexer(const std::string & s) :
Context context("When lexing dependency string '" + s + "':");
std::vector<std::string> tokens;
- Tokeniser<delim_kind::AnyOfTag, delim_mode::BoundaryTag>::tokenise(s, " \n\t", std::back_inserter(tokens));
+ tokenise<delim_kind::AnyOfTag, delim_mode::BoundaryTag>(s, " \n\t", "", std::back_inserter(tokens));
for (std::vector<std::string>::const_iterator t(tokens.begin()), t_end(tokens.end()) ;
t != t_end ; ++t)
diff --git a/paludis/repositories/e/dep_parser.cc b/paludis/repositories/e/dep_parser.cc
index e3bf49394..cd526358f 100644
--- a/paludis/repositories/e/dep_parser.cc
+++ b/paludis/repositories/e/dep_parser.cc
@@ -789,7 +789,7 @@ paludis::erepository::parse_dependency_label(const std::string & s, const EAPI &
std::set<std::string> labels;
std::string label(s.substr(0, s.length() - 1));
- Tokeniser<delim_kind::AnyOfTag>::tokenise(label, ",+", std::inserter(labels, labels.end()));
+ tokenise<delim_kind::AnyOfTag, delim_mode::DelimiterTag>(label, ",+", "", std::inserter(labels, labels.end()));
tr1::shared_ptr<DependencyLabelsDepSpec> l(new DependencyLabelsDepSpec);
diff --git a/paludis/repositories/e/e_key.cc b/paludis/repositories/e/e_key.cc
index 8a9b2ace9..876e7d699 100644
--- a/paludis/repositories/e/e_key.cc
+++ b/paludis/repositories/e/e_key.cc
@@ -564,7 +564,7 @@ EIUseKey::value() const
Context context("When parsing metadata key '" + raw_name() + "' from '" + stringify(*_imp->id) + "':");
_imp->value.reset(new IUseFlagSet);
std::list<std::string> tokens;
- WhitespaceTokeniser::tokenise(_imp->string_value, std::back_inserter(tokens));
+ tokenise_whitespace(_imp->string_value, std::back_inserter(tokens));
tr1::shared_ptr<const UseFlagNameSet> prefixes;
if (_imp->id->repository()->use_interface)
@@ -792,7 +792,7 @@ EKeywordsKey::value() const
_imp->value.reset(new KeywordNameSet);
Context context("When parsing metadata key '" + raw_name() + "' from '" + stringify(*_imp->id) + "':");
- WhitespaceTokeniser::tokenise(_imp->string_value, create_inserter<KeywordName>(_imp->value->inserter()));
+ tokenise_whitespace(_imp->string_value, create_inserter<KeywordName>(_imp->value->inserter()));
return _imp->value;
}
@@ -860,7 +860,7 @@ EUseKey::value() const
_imp->value.reset(new UseFlagNameSet);
Context context("When parsing metadata key '" + raw_name() + "' from '" + stringify(*_imp->id) + "':");
std::list<std::string> tokens;
- WhitespaceTokeniser::tokenise(_imp->string_value, std::back_inserter(tokens));
+ tokenise_whitespace(_imp->string_value, std::back_inserter(tokens));
for (std::list<std::string>::const_iterator t(tokens.begin()), t_end(tokens.end()) ;
t != t_end ; ++t)
if ('-' != t->at(0))
@@ -931,7 +931,7 @@ EInheritedKey::value() const
_imp->value.reset(new Set<std::string>);
Context context("When parsing metadata key '" + raw_name() + "' from '" + stringify(*_imp->id) + "':");
- WhitespaceTokeniser::tokenise(_imp->string_value, _imp->value->inserter());
+ tokenise_whitespace(_imp->string_value, _imp->value->inserter());
return _imp->value;
}
diff --git a/paludis/repositories/e/e_repository.cc b/paludis/repositories/e/e_repository.cc
index 269ca0f5f..604ce655a 100644
--- a/paludis/repositories/e/e_repository.cc
+++ b/paludis/repositories/e/e_repository.cc
@@ -373,8 +373,7 @@ namespace paludis
for (LineConfigFile::ConstIterator line(f.begin()), line_end(f.end()) ; line != line_end ; ++line)
{
std::vector<std::string> tokens;
- WhitespaceTokeniser::tokenise(*line,
- std::back_inserter(tokens));
+ tokenise_whitespace(*line, std::back_inserter(tokens));
if (tokens.size() < 3)
continue;
@@ -663,7 +662,7 @@ ERepository::need_mirrors() const
for (LineConfigFile::ConstIterator line(mirrors.begin()) ; line != mirrors.end() ; ++line)
{
std::vector<std::string> ee;
- WhitespaceTokeniser::tokenise(*line, std::back_inserter(ee));
+ tokenise_whitespace(*line, std::back_inserter(ee));
if (! ee.empty())
{
/* pick up to five random mirrors only */
@@ -718,7 +717,7 @@ ERepository::sync() const
return false;
std::list<std::string> sync_list;
- WhitespaceTokeniser::tokenise(_imp->params.sync, std::back_inserter(sync_list));
+ tokenise_whitespace(_imp->params.sync, std::back_inserter(sync_list));
bool ok(false);
for (std::list<std::string>::const_iterator s(sync_list.begin()),
@@ -868,7 +867,7 @@ ERepository::use_expand_flags() const
i_end(_imp->profile_ptr->end_use_expand()) ; i != i_end ; ++i)
{
std::list<std::string> values;
- WhitespaceTokeniser::tokenise(_imp->profile_ptr->environment_variable(
+ tokenise_whitespace(_imp->profile_ptr->environment_variable(
stringify(*i)), std::back_inserter(values));
for (std::list<std::string>::const_iterator j(values.begin()), j_end(values.end()) ;
j != j_end ; ++j)
diff --git a/paludis/repositories/e/e_repository_profile.cc b/paludis/repositories/e/e_repository_profile.cc
index 3a3905cf7..b62179f79 100644
--- a/paludis/repositories/e/e_repository_profile.cc
+++ b/paludis/repositories/e/e_repository_profile.cc
@@ -310,8 +310,8 @@ Implementation<ERepositoryProfile>::load_profile_make_defaults(const FSEntry & d
if (is_incremental(k->first))
{
std::list<std::string> val, val_add;
- WhitespaceTokeniser::tokenise(environment_variables[k->first], std::back_inserter(val));
- WhitespaceTokeniser::tokenise(k->second, std::back_inserter(val_add));
+ tokenise_whitespace(environment_variables[k->first], std::back_inserter(val));
+ tokenise_whitespace(k->second, std::back_inserter(val_add));
for (std::list<std::string>::const_iterator v(val_add.begin()), v_end(val_add.end()) ;
v != v_end ; ++v)
@@ -338,7 +338,7 @@ Implementation<ERepositoryProfile>::load_profile_make_defaults(const FSEntry & d
{
use_expand.clear();
if (! use_expand_var.empty())
- WhitespaceTokeniser::tokenise(environment_variables[use_expand_var],
+ tokenise_whitespace(environment_variables[use_expand_var],
create_inserter<UseFlagName>(std::inserter(use_expand, use_expand.end())));
}
catch (const Exception & e)
@@ -357,7 +357,7 @@ Implementation<ERepositoryProfile>::load_special_make_defaults_vars()
{
use.clear();
if (! use_var.empty())
- WhitespaceTokeniser::tokenise(environment_variables[use_var],
+ tokenise_whitespace(environment_variables[use_var],
create_inserter<UseFlagName>(std::inserter(use, use.end())));
}
catch (const Exception & e)
@@ -372,7 +372,7 @@ Implementation<ERepositoryProfile>::load_special_make_defaults_vars()
{
use_expand.clear();
if (! use_expand_var.empty())
- WhitespaceTokeniser::tokenise(environment_variables[use_expand_var],
+ tokenise_whitespace(environment_variables[use_expand_var],
create_inserter<UseFlagName>(std::inserter(use_expand, use_expand.end())));
}
catch (const Exception & e)
@@ -387,7 +387,7 @@ Implementation<ERepositoryProfile>::load_special_make_defaults_vars()
{
use_expand_hidden.clear();
if (! use_expand_hidden_var.empty())
- WhitespaceTokeniser::tokenise(environment_variables[use_expand_hidden_var],
+ tokenise_whitespace(environment_variables[use_expand_hidden_var],
create_inserter<UseFlagName>(std::inserter(use_expand_hidden, use_expand_hidden.end())));
}
catch (const Exception & e)
@@ -460,7 +460,7 @@ Implementation<ERepositoryProfile>::make_vars_from_file_vars()
line != line_end ; ++line)
{
std::vector<std::string> tokens;
- WhitespaceTokeniser::tokenise(*line, std::back_inserter(tokens));
+ tokenise_whitespace(*line, std::back_inserter(tokens));
if (tokens.size() < 2)
continue;
@@ -513,7 +513,7 @@ Implementation<ERepositoryProfile>::load_basic_use_file(const FSEntry & file, Fl
line != line_end ; ++line)
{
std::list<std::string> tokens;
- WhitespaceTokeniser::tokenise(*line, std::back_inserter(tokens));
+ tokenise_whitespace(*line, std::back_inserter(tokens));
for (std::list<std::string>::const_iterator t(tokens.begin()), t_end(tokens.end()) ;
t != t_end ; ++t)
@@ -548,7 +548,7 @@ Implementation<ERepositoryProfile>::load_spec_use_file(const FSEntry & file, Pac
line != line_end ; ++line)
{
std::list<std::string> tokens;
- WhitespaceTokeniser::tokenise(*line, std::back_inserter(tokens));
+ tokenise_whitespace(*line, std::back_inserter(tokens));
if (tokens.empty())
continue;
@@ -603,7 +603,7 @@ Implementation<ERepositoryProfile>::add_use_expand_to_use()
&::tolower);
std::list<std::string> uses;
- WhitespaceTokeniser::tokenise(environment_variables[stringify(*x)],
+ tokenise_whitespace(environment_variables[stringify(*x)],
std::back_inserter(uses));
for (std::list<std::string>::const_iterator u(uses.begin()), u_end(uses.end()) ;
u != u_end ; ++u)
diff --git a/paludis/repositories/e/eapi.cc b/paludis/repositories/e/eapi.cc
index db5a9014b..b2a933f44 100644
--- a/paludis/repositories/e/eapi.cc
+++ b/paludis/repositories/e/eapi.cc
@@ -176,11 +176,11 @@ namespace paludis
))));
- WhitespaceTokeniser::tokenise(k.get("restrict_fetch"),
+ tokenise_whitespace(k.get("restrict_fetch"),
eapi->supported->ebuild_options->restrict_fetch->inserter());
- WhitespaceTokeniser::tokenise(k.get("restrict_mirror"),
+ tokenise_whitespace(k.get("restrict_mirror"),
eapi->supported->ebuild_options->restrict_mirror->inserter());
- WhitespaceTokeniser::tokenise(k.get("restrict_primaryuri"),
+ tokenise_whitespace(k.get("restrict_primaryuri"),
eapi->supported->ebuild_options->restrict_primaryuri->inserter());
values.insert(std::make_pair(strip_trailing_string(d->basename(), ".conf"), eapi));
@@ -239,13 +239,13 @@ EAPILabels::EAPILabels(const std::string & s) :
{
std::vector<std::string> tokens;
- Tokeniser<delim_kind::AnyOfTag>::tokenise(s, ";", std::back_inserter(tokens));
+ tokenise<delim_kind::AnyOfTag, delim_mode::DelimiterTag>(s, ";", "", std::back_inserter(tokens));
for (std::vector<std::string>::const_iterator t(tokens.begin()), t_end(tokens.end()) ;
t != t_end ; ++t)
{
std::vector<std::string> values;
- Tokeniser<delim_kind::AnyOfTag>::tokenise(*t, "=", std::back_inserter(values));
+ tokenise<delim_kind::AnyOfTag, delim_mode::DelimiterTag>(*t, "=", "", std::back_inserter(values));
if (values.size() != 2)
throw EAPIConfigurationError("EAPI labels value '" + s + "' has bad values size '" + stringify(values.size()) + "'");
diff --git a/paludis/repositories/e/eapi_phase.cc b/paludis/repositories/e/eapi_phase.cc
index 936dae706..32a40d8bc 100644
--- a/paludis/repositories/e/eapi_phase.cc
+++ b/paludis/repositories/e/eapi_phase.cc
@@ -57,7 +57,7 @@ EAPIPhase::EAPIPhase(const std::string & s) :
Context c("When parsing EAPI phase '" + s + "'");
std::list<std::string> tokens;
- WhitespaceTokeniser::tokenise(s, std::back_inserter(tokens));
+ tokenise_whitespace(s, std::back_inserter(tokens));
std::list<std::string>::iterator t(std::find(tokens.begin(), tokens.end(), ":"));
if (t == tokens.end())
@@ -95,7 +95,7 @@ EAPIPhases::EAPIPhases(const std::string & s) :
Context c("When parsing EAPI phases '" + s + "'");
std::list<std::string> tokens;
- Tokeniser<delim_kind::AnyOfTag>::tokenise(s, ";", std::back_inserter(tokens));
+ tokenise<delim_kind::AnyOfTag, delim_mode::DelimiterTag>(s, ";", "", std::back_inserter(tokens));
for (std::list<std::string>::const_iterator t(tokens.begin()), t_end(tokens.end()) ;
t != t_end ; ++t)
_imp->phases.push_back(make_shared_ptr(new EAPIPhase(*t)));
diff --git a/paludis/repositories/e/ebuild.cc b/paludis/repositories/e/ebuild.cc
index 8715eacfe..f84b03673 100644
--- a/paludis/repositories/e/ebuild.cc
+++ b/paludis/repositories/e/ebuild.cc
@@ -253,7 +253,7 @@ namespace
std::string purdy(const std::string & s)
{
std::list<std::string> tokens;
- WhitespaceTokeniser::tokenise(s, std::back_inserter(tokens));
+ tokenise_whitespace(s, std::back_inserter(tokens));
return join(tokens.begin(), tokens.end(), " \\n ");
}
}
diff --git a/paludis/repositories/e/ebuild_entries.cc b/paludis/repositories/e/ebuild_entries.cc
index bafed93b5..13f06012e 100644
--- a/paludis/repositories/e/ebuild_entries.cc
+++ b/paludis/repositories/e/ebuild_entries.cc
@@ -221,7 +221,7 @@ namespace
/* possible values from profile */
std::set<UseFlagName> possible_values;
- WhitespaceTokeniser::tokenise(profile->environment_variable(stringify(*x)),
+ tokenise_whitespace(profile->environment_variable(stringify(*x)),
create_inserter<UseFlagName>(std::inserter(possible_values, possible_values.end())));
/* possible values from environment */
diff --git a/paludis/repositories/e/ebuild_flat_metadata_cache.cc b/paludis/repositories/e/ebuild_flat_metadata_cache.cc
index 824edc2fe..de564f228 100644
--- a/paludis/repositories/e/ebuild_flat_metadata_cache.cc
+++ b/paludis/repositories/e/ebuild_flat_metadata_cache.cc
@@ -77,7 +77,7 @@ EbuildFlatMetadataCache::load(const tr1::shared_ptr<const EbuildID> & id)
{
time_t cache_time(std::max(_master_mtime, _filename.mtime()));
std::set<std::string> tokens;
- WhitespaceTokeniser::tokenise(lines[9], std::inserter(tokens, tokens.begin()));
+ tokenise_whitespace(lines[9], std::inserter(tokens, tokens.begin()));
ok = _ebuild.mtime() <= cache_time;
if (ok && ! tokens.empty())
@@ -152,7 +152,7 @@ namespace
std::string normalise(const T_ & s)
{
std::list<std::string> tokens;
- WhitespaceTokeniser::tokenise(stringify(s), std::back_inserter(tokens));
+ tokenise_whitespace(stringify(s), std::back_inserter(tokens));
return join(tokens.begin(), tokens.end(), " ");
}
diff --git a/paludis/repositories/e/make_ebin_repository.cc b/paludis/repositories/e/make_ebin_repository.cc
index dfd0ce6b1..682f44503 100644
--- a/paludis/repositories/e/make_ebin_repository.cc
+++ b/paludis/repositories/e/make_ebin_repository.cc
@@ -76,7 +76,7 @@ paludis::make_ebin_repository(
tr1::shared_ptr<FSEntrySequence> profiles(new FSEntrySequence);
if (m->end() != m->find("profiles"))
- WhitespaceTokeniser::tokenise(m->find("profiles")->second,
+ tokenise_whitespace(m->find("profiles")->second,
create_inserter<FSEntry>(std::back_inserter(*profiles)));
if (profiles->empty())
diff --git a/paludis/repositories/e/make_ebuild_repository.cc b/paludis/repositories/e/make_ebuild_repository.cc
index 5f50796f5..247e856cd 100644
--- a/paludis/repositories/e/make_ebuild_repository.cc
+++ b/paludis/repositories/e/make_ebuild_repository.cc
@@ -76,7 +76,7 @@ paludis::make_ebuild_repository(
tr1::shared_ptr<FSEntrySequence> profiles(new FSEntrySequence);
if (m->end() != m->find("profiles"))
- WhitespaceTokeniser::tokenise(m->find("profiles")->second,
+ tokenise_whitespace(m->find("profiles")->second,
create_inserter<FSEntry>(std::back_inserter(*profiles)));
if (profiles->empty())
@@ -91,7 +91,7 @@ paludis::make_ebuild_repository(
tr1::shared_ptr<FSEntrySequence> eclassdirs(new FSEntrySequence);
if (m->end() != m->find("eclassdirs"))
- WhitespaceTokeniser::tokenise(m->find("eclassdirs")->second,
+ tokenise_whitespace(m->find("eclassdirs")->second,
create_inserter<FSEntry>(std::back_inserter(*eclassdirs)));
if (eclassdirs->empty())
diff --git a/paludis/repositories/e/manifest2_reader.cc b/paludis/repositories/e/manifest2_reader.cc
index 5a2b09f25..9446a78b8 100644
--- a/paludis/repositories/e/manifest2_reader.cc
+++ b/paludis/repositories/e/manifest2_reader.cc
@@ -72,8 +72,7 @@ Manifest2Reader::Manifest2Reader(const FSEntry & f) :
l != l_end ; ++l)
{
std::list<std::string> tokens;
- WhitespaceTokeniser::tokenise((*l),
- create_inserter<std::string>(std::back_inserter(tokens)));
+ tokenise_whitespace((*l), create_inserter<std::string>(std::back_inserter(tokens)));
std::list<std::string>::const_iterator t(tokens.begin()), t_end(tokens.end());
std::string type, name, sha1, sha256, rmd160, md5;
diff --git a/paludis/repositories/e/qa/kv_variables.cc b/paludis/repositories/e/qa/kv_variables.cc
index 3e2218e94..912575606 100644
--- a/paludis/repositories/e/qa/kv_variables.cc
+++ b/paludis/repositories/e/qa/kv_variables.cc
@@ -44,7 +44,7 @@ paludis::erepository::kv_variables_check(
<< entry << "', '" << *id << "', '" << name << "'";
std::list<std::string> lines;
- Tokeniser<delim_kind::AnyOfTag>::tokenise(content, "\n", std::back_inserter(lines));
+ tokenise<delim_kind::AnyOfTag, delim_mode::DelimiterTag>(content, "\n", "", std::back_inserter(lines));
unsigned line(0);
for (std::list<std::string>::const_iterator l(lines.begin()), l_end(lines.end()) ;
diff --git a/paludis/repositories/e/qa/visibility.cc b/paludis/repositories/e/qa/visibility.cc
index 6fe517bce..0df42ad3f 100644
--- a/paludis/repositories/e/qa/visibility.cc
+++ b/paludis/repositories/e/qa/visibility.cc
@@ -288,7 +288,7 @@ paludis::erepository::visibility_check(
continue;
std::set<KeywordName> accepted_keywords, overlap;
- WhitespaceTokeniser::tokenise(p->profile->environment_variable(
+ tokenise_whitespace(p->profile->environment_variable(
repo->accept_keywords_variable()), create_inserter<KeywordName>(std::inserter(accepted_keywords, accepted_keywords.begin())));
std::set_intersection(accepted_keywords.begin(), accepted_keywords.end(),
diff --git a/paludis/repositories/e/qa/whitespace.cc b/paludis/repositories/e/qa/whitespace.cc
index 705c0aef0..4688ee792 100644
--- a/paludis/repositories/e/qa/whitespace.cc
+++ b/paludis/repositories/e/qa/whitespace.cc
@@ -40,7 +40,7 @@ paludis::erepository::whitespace_check(
<< entry << "', '" << *id << "', '" << name << "'";
std::list<std::string> lines;
- Tokeniser<delim_kind::AnyOfTag>::tokenise(content, "\n", std::back_inserter(lines));
+ tokenise<delim_kind::AnyOfTag, delim_mode::DelimiterTag>(content, "\n", "", std::back_inserter(lines));
unsigned line(0), err_count(0);
for (std::list<std::string>::const_iterator l(lines.begin()), l_end(lines.end()) ;
diff --git a/paludis/repositories/e/vdb_merger.cc b/paludis/repositories/e/vdb_merger.cc
index 3a714b047..24ddf5428 100644
--- a/paludis/repositories/e/vdb_merger.cc
+++ b/paludis/repositories/e/vdb_merger.cc
@@ -57,10 +57,8 @@ namespace paludis
options(o),
realroot(options.root.realpath())
{
- WhitespaceTokeniser::tokenise(o.config_protect,
- std::back_inserter(config_protect));
- WhitespaceTokeniser::tokenise(o.config_protect_mask,
- std::back_inserter(config_protect_mask));
+ tokenise_whitespace(o.config_protect, std::back_inserter(config_protect));
+ tokenise_whitespace(o.config_protect_mask, std::back_inserter(config_protect_mask));
}
};
}
diff --git a/paludis/repositories/e/vdb_repository.cc b/paludis/repositories/e/vdb_repository.cc
index 5a9d97b30..913b6cfde 100644
--- a/paludis/repositories/e/vdb_repository.cc
+++ b/paludis/repositories/e/vdb_repository.cc
@@ -871,7 +871,7 @@ VDBRepository::load_provided_using_cache() const
while (std::getline(provides_cache, line))
{
std::vector<std::string> tokens;
- WhitespaceTokeniser::tokenise(line, std::back_inserter(tokens));
+ tokenise_whitespace(line, std::back_inserter(tokens));
if (tokens.size() < 3)
continue;
diff --git a/paludis/repositories/e/vdb_unmerger.cc b/paludis/repositories/e/vdb_unmerger.cc
index 043f4ccd9..209970b32 100644
--- a/paludis/repositories/e/vdb_unmerger.cc
+++ b/paludis/repositories/e/vdb_unmerger.cc
@@ -58,10 +58,8 @@ namespace paludis
Implementation(const VDBUnmergerOptions & o) :
options(o)
{
- WhitespaceTokeniser::tokenise(o.config_protect,
- std::back_inserter(config_protect));
- WhitespaceTokeniser::tokenise(o.config_protect_mask,
- std::back_inserter(config_protect_mask));
+ tokenise_whitespace(o.config_protect, std::back_inserter(config_protect));
+ tokenise_whitespace(o.config_protect_mask, std::back_inserter(config_protect_mask));
}
};
}
diff --git a/paludis/repositories/e/xml_things.cc b/paludis/repositories/e/xml_things.cc
index fcfb4d63e..e4a75d30b 100644
--- a/paludis/repositories/e/xml_things.cc
+++ b/paludis/repositories/e/xml_things.cc
@@ -43,7 +43,7 @@ namespace
std::string normalise(const std::string & s)
{
std::list<std::string> words;
- WhitespaceTokeniser::tokenise(s, std::back_inserter(words));
+ tokenise_whitespace(s, std::back_inserter(words));
return join(words.begin(), words.end(), " ");
}
@@ -96,8 +96,7 @@ namespace
if (name == "arch")
{
std::set<std::string> archs;
- WhitespaceTokeniser::tokenise(retarded_libxml_string_to_string(
- xmlNodeListGetString(doc, a->xmlChildrenNode, 1)),
+ tokenise_whitespace(retarded_libxml_string_to_string(xmlNodeListGetString(doc, a->xmlChildrenNode, 1)),
std::inserter(archs, archs.end()));
archs.erase("*");
for (std::set<std::string>::const_iterator r(archs.begin()), r_end(archs.end()) ;
diff --git a/paludis/repositories/fake/fake_package_id.cc b/paludis/repositories/fake/fake_package_id.cc
index adef75c0e..ede593c0d 100644
--- a/paludis/repositories/fake/fake_package_id.cc
+++ b/paludis/repositories/fake/fake_package_id.cc
@@ -98,7 +98,7 @@ void
FakeMetadataKeywordSetKey::set_from_string(const std::string & s)
{
_imp->collection.reset(new KeywordNameSet);
- WhitespaceTokeniser::tokenise(s, create_inserter<KeywordName>(_imp->collection->inserter()));
+ tokenise_whitespace(s, create_inserter<KeywordName>(_imp->collection->inserter()));
}
FakeMetadataIUseSetKey::FakeMetadataIUseSetKey(const std::string & r,
@@ -114,7 +114,7 @@ FakeMetadataIUseSetKey::set_from_string(const std::string & s, const IUseFlagPar
{
_imp->collection.reset(new IUseFlagSet);
std::list<std::string> tokens;
- WhitespaceTokeniser::tokenise(s, std::back_inserter(tokens));
+ tokenise_whitespace(s, std::back_inserter(tokens));
for (std::list<std::string>::const_iterator t(tokens.begin()), t_end(tokens.end()) ;
t != t_end ; ++t)
_imp->collection->insert(IUseFlag(*t, m, std::string::npos));
diff --git a/paludis/repositories/unpackaged/dep_parser.cc b/paludis/repositories/unpackaged/dep_parser.cc
index b5fe2b392..8416d29a5 100644
--- a/paludis/repositories/unpackaged/dep_parser.cc
+++ b/paludis/repositories/unpackaged/dep_parser.cc
@@ -36,7 +36,7 @@ DepParser::parse(const std::string & s)
new ConstTreeSequence<DependencySpecTree, AllDepSpec>(tr1::shared_ptr<AllDepSpec>(new AllDepSpec)));
std::list<std::string> tokens;
- Tokeniser<delim_kind::AnyOfTag>::tokenise(s, ",", std::back_inserter(tokens));
+ tokenise<delim_kind::AnyOfTag, delim_mode::BoundaryTag>(s, ",", "", std::back_inserter(tokens));
for (std::list<std::string>::const_iterator t(tokens.begin()), t_end(tokens.end()) ;
t != t_end ; ++t)
diff --git a/paludis/repositories/unpackaged/ndbam.cc b/paludis/repositories/unpackaged/ndbam.cc
index 027642262..2a2a73862 100644
--- a/paludis/repositories/unpackaged/ndbam.cc
+++ b/paludis/repositories/unpackaged/ndbam.cc
@@ -299,7 +299,7 @@ NDBAM::entries(const QualifiedPackageName & q)
try
{
std::vector<std::string> tokens;
- Tokeniser<delim_kind::AnyOfTag>::tokenise(d->basename(), ":", std::back_inserter(tokens));
+ tokenise<delim_kind::AnyOfTag, delim_mode::DelimiterTag>(d->basename(), ":", "", std::back_inserter(tokens));
if (tokens.size() < 3)
{
Log::get_instance()->message(ll_warning, lc_context) << "Not using '" << *d <<
diff --git a/paludis/repositories/unpackaged/ndbam_merger.cc b/paludis/repositories/unpackaged/ndbam_merger.cc
index 5674094d5..d0a2ac3e0 100644
--- a/paludis/repositories/unpackaged/ndbam_merger.cc
+++ b/paludis/repositories/unpackaged/ndbam_merger.cc
@@ -58,10 +58,8 @@ namespace paludis
options(o),
realroot(options.root.realpath())
{
- WhitespaceTokeniser::tokenise(o.config_protect,
- std::back_inserter(config_protect));
- WhitespaceTokeniser::tokenise(o.config_protect_mask,
- std::back_inserter(config_protect_mask));
+ tokenise_whitespace(o.config_protect, std::back_inserter(config_protect));
+ tokenise_whitespace(o.config_protect_mask, std::back_inserter(config_protect_mask));
}
};
}
diff --git a/paludis/repositories/unpackaged/ndbam_unmerger.cc b/paludis/repositories/unpackaged/ndbam_unmerger.cc
index 80f598260..5f495160f 100644
--- a/paludis/repositories/unpackaged/ndbam_unmerger.cc
+++ b/paludis/repositories/unpackaged/ndbam_unmerger.cc
@@ -62,10 +62,8 @@ namespace paludis
Implementation(const NDBAMUnmergerOptions & o) :
options(o)
{
- WhitespaceTokeniser::tokenise(o.config_protect,
- std::back_inserter(config_protect));
- WhitespaceTokeniser::tokenise(o.config_protect_mask,
- std::back_inserter(config_protect_mask));
+ tokenise_whitespace(o.config_protect, std::back_inserter(config_protect));
+ tokenise_whitespace(o.config_protect_mask, std::back_inserter(config_protect_mask));
}
};
}