opennlp.tools.util.featuregen.BrownCluster Maven / Gradle / Ivy
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package opennlp.tools.util.featuregen;
import java.io.BufferedReader;
import java.io.BufferedWriter;
import java.io.IOException;
import java.io.InputStream;
import java.io.InputStreamReader;
import java.io.OutputStream;
import java.io.OutputStreamWriter;
import java.io.Writer;
import java.nio.charset.StandardCharsets;
import java.util.HashMap;
import java.util.Map;
import java.util.regex.Pattern;
import opennlp.tools.util.model.ArtifactSerializer;
import opennlp.tools.util.model.SerializableArtifact;
/**
*
* Class to load a Brown cluster document: word\tword_class\tprob
*
* Originally available at:
* http://metaoptimize.com/projects/wordreprs/.
*
* Further details can be found in the
* related research paper.
*
* The file containing the clustering lexicon has to be passed as the
* value of the dict attribute of each {@link BrownCluster} feature generator.
*/
public class BrownCluster implements SerializableArtifact {
private static final Pattern tabPattern = Pattern.compile("\t");
public static class BrownClusterSerializer implements ArtifactSerializer {
@Override
public BrownCluster create(InputStream in) throws IOException {
return new BrownCluster(in);
}
@Override
public void serialize(BrownCluster artifact, OutputStream out)
throws IOException {
artifact.serialize(out);
}
}
private final Map tokenToClusterMap = new HashMap<>();
/**
* Instatiates a {@link BrownCluster} and its related token to cluster map
* via an {@link InputStream}.
*
* Note:
* Only tokens with frequency bigger than {@code 5} will be added.
*
* @param in A valid, open {@link InputStream} to read from.
* @throws IOException Thrown if errors occurred reading from {@link InputStream in}.
*/
public BrownCluster(InputStream in) throws IOException {
try (BufferedReader breader = new BufferedReader(
new InputStreamReader(in, StandardCharsets.UTF_8))) {
String line;
while ((line = breader.readLine()) != null) {
String[] lineArray = tabPattern.split(line);
if (lineArray.length == 3) {
int freq = Integer.parseInt(lineArray[2]);
if (freq > 5 ) {
tokenToClusterMap.put(lineArray[1], lineArray[0]);
}
}
else if (lineArray.length == 2) {
tokenToClusterMap.put(lineArray[0], lineArray[1]);
}
}
}
}
/**
* Check if a token is in the Brown:paths, token map.
*
* @param string the token to look-up
* @return the brown class if such token is in the brown cluster map
*/
public String lookupToken(String string) {
return tokenToClusterMap.get(string);
}
public void serialize(OutputStream out) throws IOException {
try (Writer writer = new BufferedWriter(new OutputStreamWriter(out))) {
for (Map.Entry entry : tokenToClusterMap.entrySet()) {
writer.write(entry.getKey() + "\t" + entry.getValue() + "\n");
}
writer.flush();
}
}
@Override
public Class> getArtifactSerializerClass() {
return BrownClusterSerializer.class;
}
}