All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.tika.parser.txt.CharsetDetector Maven / Gradle / Ivy

There is a newer version: 3.0.0-BETA2
Show newest version
/**
 * ******************************************************************************
 * Copyright (C) 2005-2009, International Business Machines Corporation and    *
 * others. All Rights Reserved.                                                *
 * ******************************************************************************
 */
package org.apache.tika.parser.txt;

import java.io.IOException;
import java.io.InputStream;
import java.io.Reader;
import java.nio.charset.Charset;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.Collections;


/**
 * CharsetDetector provides a facility for detecting the
 * charset or encoding of character data in an unknown format.
 * The input data can either be from an input stream or an array of bytes.
 * The result of the detection operation is a list of possibly matching
 * charsets, or, for simple use, you can just ask for a Java Reader that
 * will will work over the input data.
 * 

* Character set detection is at best an imprecise operation. The detection * process will attempt to identify the charset that best matches the characteristics * of the byte data, but the process is partly statistical in nature, and * the results can not be guaranteed to always be correct. *

* For best accuracy in charset detection, the input data should be primarily * in a single language, and a minimum of a few hundred bytes worth of plain text * in the language are needed. The detection process will attempt to * ignore html or xml style markup that could otherwise obscure the content. *

* @stable ICU 3.4 */ public class CharsetDetector { // Question: Should we have getters corresponding to the setters for input text // and declared encoding? // A thought: If we were to create our own type of Java Reader, we could defer // figuring out an actual charset for data that starts out with too much English // only ASCII until the user actually read through to something that didn't look // like 7 bit English. If nothing else ever appeared, we would never need to // actually choose the "real" charset. All assuming that the application just // wants the data, and doesn't care about a char set name. private static final int kBufSize = 12000; private static final int MAX_CONFIDENCE = 100; private static String[] fCharsetNames; /* * List of recognizers for all charsets known to the implementation. */ private static ArrayList fCSRecognizers = createRecognizers(); /* * The following items are accessed by individual CharsetRecongizers during * the recognition process * */ byte[] fInputBytes = // The text to be checked. Markup will have been new byte[kBufSize]; // removed if appropriate. int fInputLen; // Length of the byte data in fInputText. short fByteStats[] = // byte frequency statistics for the input text. new short[256]; // Value is percent, not absolute. boolean fC1Bytes = // True if any bytes in the range 0x80 - 0x9F are in the input; false; String fDeclaredEncoding; // // Stuff private to CharsetDetector // byte[] fRawInput; // Original, untouched input bytes. // If user gave us a byte array, this is it. // If user gave us a stream, it's read to a // buffer here. int fRawLength; // Length of data in fRawInput array. InputStream fInputStream; // User's input stream, or null if the user boolean fStripTags = // If true, setText() will strip tags from input text. false; /** * Constructor * * @stable ICU 3.4 */ public CharsetDetector() { } /** * Get the names of all char sets that can be recognized by the char set detector. * * @return an array of the names of all charsets that can be recognized * by the charset detector. * * @stable ICU 3.4 */ public static String[] getAllDetectableCharsets() { return fCharsetNames; } /* * Create the singleton instances of the CharsetRecognizer classes */ private static ArrayList createRecognizers() { ArrayList recognizers = new ArrayList(); recognizers.add(new CharsetRecog_UTF8()); recognizers.add(new CharsetRecog_Unicode.CharsetRecog_UTF_16_BE()); recognizers.add(new CharsetRecog_Unicode.CharsetRecog_UTF_16_LE()); recognizers.add(new CharsetRecog_Unicode.CharsetRecog_UTF_32_BE()); recognizers.add(new CharsetRecog_Unicode.CharsetRecog_UTF_32_LE()); recognizers.add(new CharsetRecog_mbcs.CharsetRecog_sjis()); recognizers.add(new CharsetRecog_2022.CharsetRecog_2022JP()); recognizers.add(new CharsetRecog_2022.CharsetRecog_2022CN()); recognizers.add(new CharsetRecog_2022.CharsetRecog_2022KR()); recognizers.add(new CharsetRecog_mbcs.CharsetRecog_euc.CharsetRecog_gb_18030()); recognizers.add(new CharsetRecog_mbcs.CharsetRecog_euc.CharsetRecog_euc_jp()); recognizers.add(new CharsetRecog_mbcs.CharsetRecog_euc.CharsetRecog_euc_kr()); recognizers.add(new CharsetRecog_mbcs.CharsetRecog_big5()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_8859_1_da()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_8859_1_de()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_8859_1_en()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_8859_1_es()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_8859_1_fr()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_8859_1_it()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_8859_1_nl()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_8859_1_no()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_8859_1_pt()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_8859_1_sv()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_8859_2_cs()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_8859_2_hu()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_8859_2_pl()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_8859_2_ro()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_8859_5_ru()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_8859_6_ar()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_8859_7_el()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_8859_8_I_he()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_8859_8_he()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_windows_1251()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_windows_1256()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_KOI8_R()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_8859_9_tr()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_IBM424_he_rtl()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_IBM424_he_ltr()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_IBM420_ar_rtl()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_IBM420_ar_ltr()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_EBCDIC_500_en()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_EBCDIC_500_de()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_EBCDIC_500_es()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_EBCDIC_500_fr()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_EBCDIC_500_it()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_EBCDIC_500_nl()); recognizers.add(new CharsetRecog_sbcs.CharsetRecog_IBM866_ru()); // Create an array of all charset names, as a side effect. // Needed for the getAllDetectableCharsets() API. String[] charsetNames = new String[recognizers.size()]; int out = 0; for (CharsetRecognizer recognizer : recognizers) { String name = recognizer.getName(); if (out == 0 || !name.equals(charsetNames[out - 1])) { charsetNames[out++] = name; } } fCharsetNames = new String[out]; System.arraycopy(charsetNames, 0, fCharsetNames, 0, out); return recognizers; } /** * Set the declared encoding for charset detection. * The declared encoding of an input text is an encoding obtained * from an http header or xml declaration or similar source that * can be provided as additional information to the charset detector. * A match between a declared encoding and a possible detected encoding * will raise the quality of that detected encoding by a small delta, * and will also appear as a "reason" for the match. *

* A declared encoding that is incompatible with the input data being * analyzed will not be added to the list of possible encodings. * * @param encoding The declared encoding * * @stable ICU 3.4 */ public CharsetDetector setDeclaredEncoding(String encoding) { setCanonicalDeclaredEncoding(encoding); return this; } /** * Set the input text (byte) data whose charset is to be detected. * * @param in the input text of unknown encoding * * @return This CharsetDetector * * @stable ICU 3.4 */ public CharsetDetector setText(byte[] in) { fRawInput = in; fRawLength = in.length; MungeInput(); return this; } // Value is rounded up, so zero really means zero occurences. /** * Set the input text (byte) data whose charset is to be detected. *

* The input stream that supplies the character data must have markSupported() * == true; the charset detection process will read a small amount of data, * then return the stream to its original position via * the InputStream.reset() operation. The exact amount that will * be read depends on the characteristics of the data itself. * * @param in the input text of unknown encoding * * @return This CharsetDetector * * @stable ICU 3.4 */ public CharsetDetector setText(InputStream in) throws IOException { fInputStream = in; fInputStream.mark(kBufSize); fRawInput = new byte[kBufSize]; // Always make a new buffer because the // previous one may have come from the caller, // in which case we can't touch it. fRawLength = 0; int remainingLength = kBufSize; while (remainingLength > 0) { // read() may give data in smallish chunks, esp. for remote sources. Hence, this loop. int bytesRead = fInputStream.read(fRawInput, fRawLength, remainingLength); if (bytesRead <= 0) { break; } fRawLength += bytesRead; remainingLength -= bytesRead; } fInputStream.reset(); MungeInput(); // Strip html markup, collect byte stats. return this; } /** * Return the charset that best matches the supplied input data. * * Note though, that because the detection * only looks at the start of the input data, * there is a possibility that the returned charset will fail to handle * the full set of input data. *

* Raise an exception if *

    *
  • no charset appears to match the data.
  • *
  • no input text has been provided
  • *
* * @return a CharsetMatch object representing the best matching charset, or * null if there are no matches. * * @stable ICU 3.4 */ public CharsetMatch detect() { // TODO: A better implementation would be to copy the detect loop from // detectAll(), and cut it short as soon as a match with a high confidence // is found. This is something to be done later, after things are otherwise // working. CharsetMatch matches[] = detectAll(); if (matches == null || matches.length == 0) { return null; } return matches[0]; } /** * Return an array of all charsets that appear to be plausible * matches with the input data. The array is ordered with the * best quality match first. *

* Raise an exception if *

    *
  • no charsets appear to match the input data.
  • *
  • no input text has been provided
  • *
* * @return An array of CharsetMatch objects representing possibly matching charsets. * * @stable ICU 3.4 */ public CharsetMatch[] detectAll() { CharsetRecognizer csr; int i; int detectResults; int confidence; ArrayList matches = new ArrayList(); // Iterate over all possible charsets, remember all that // give a match quality > 0. for (i = 0; i < fCSRecognizers.size(); i++) { csr = fCSRecognizers.get(i); detectResults = csr.match(this); confidence = detectResults & 0x000000ff; if (confidence > 0) { // Just to be safe, constrain confidence = Math.min(confidence, MAX_CONFIDENCE); // Apply charset hint. if ((fDeclaredEncoding != null) && (fDeclaredEncoding.equalsIgnoreCase(csr.getName()))) { // Reduce lack of confidence (delta between "sure" and current) by 50%. confidence += (MAX_CONFIDENCE - confidence) / 2; } CharsetMatch m = new CharsetMatch(this, csr, confidence); matches.add(m); } } Collections.sort(matches); // CharsetMatch compares on confidence Collections.reverse(matches); // Put best match first. CharsetMatch[] resultArray = new CharsetMatch[matches.size()]; resultArray = matches.toArray(resultArray); return resultArray; } /** * Autodetect the charset of an inputStream, and return a Java Reader * to access the converted input data. *

* This is a convenience method that is equivalent to * this.setDeclaredEncoding(declaredEncoding).setText(in).detect().getReader(); *

* For the input stream that supplies the character data, markSupported() * must be true; the charset detection will read a small amount of data, * then return the stream to its original position via * the InputStream.reset() operation. The exact amount that will * be read depends on the characteristics of the data itself. *

* Raise an exception if no charsets appear to match the input data. * * @param in The source of the byte data in the unknown charset. * * @param declaredEncoding A declared encoding for the data, if available, * or null or an empty string if none is available. * * @stable ICU 3.4 */ public Reader getReader(InputStream in, String declaredEncoding) { setCanonicalDeclaredEncoding(declaredEncoding); try { setText(in); CharsetMatch match = detect(); if (match == null) { return null; } return match.getReader(); } catch (IOException e) { return null; } } /** * Autodetect the charset of an inputStream, and return a String * containing the converted input data. *

* This is a convenience method that is equivalent to * this.setDeclaredEncoding(declaredEncoding).setText(in).detect().getString(); *

* Raise an exception if no charsets appear to match the input data. * * @param in The source of the byte data in the unknown charset. * * @param declaredEncoding A declared encoding for the data, if available, * or null or an empty string if none is available. * * @stable ICU 3.4 */ public String getString(byte[] in, String declaredEncoding) { setCanonicalDeclaredEncoding(declaredEncoding); try { setText(in); CharsetMatch match = detect(); if (match == null) { return null; } return match.getString(-1); } catch (IOException e) { return null; } } // gave us a byte array. /** * Test whether or not input filtering is enabled. * * @return true if input text will be filtered. * * @see #enableInputFilter * * @stable ICU 3.4 */ public boolean inputFilterEnabled() { return fStripTags; } /** * Enable filtering of input text. If filtering is enabled, * text within angle brackets ("<" and ">") will be removed * before detection. * * @param filter true to enable input text filtering. * * @return The previous setting. * * @stable ICU 3.4 */ public boolean enableInputFilter(boolean filter) { boolean previous = fStripTags; fStripTags = filter; return previous; } /** * Try to set fDeclaredEncoding to the canonical name for , if it exists. * * @param encoding - name of character encoding */ private void setCanonicalDeclaredEncoding(String encoding) { if ((encoding == null) || encoding.isEmpty()) { return; } Charset cs = Charset.forName(encoding); if (cs != null) { fDeclaredEncoding = cs.name(); } } /* * MungeInput - after getting a set of raw input data to be analyzed, preprocess * it by removing what appears to be html markup. */ private void MungeInput() { int srci = 0; int dsti = 0; byte b; boolean inMarkup = false; int openTags = 0; int badTags = 0; // // html / xml markup stripping. // quick and dirty, not 100% accurate, but hopefully good enough, statistically. // discard everything within < brackets > // Count how many total '<' and illegal (nested) '<' occur, so we can make some // guess as to whether the input was actually marked up at all. if (fStripTags) { for (srci = 0; srci < fRawLength && dsti < fInputBytes.length; srci++) { b = fRawInput[srci]; if (b == (byte) '<') { if (inMarkup) { badTags++; } inMarkup = true; openTags++; } if (!inMarkup) { fInputBytes[dsti++] = b; } if (b == (byte) '>') { inMarkup = false; } } fInputLen = dsti; } // // If it looks like this input wasn't marked up, or if it looks like it's // essentially nothing but markup abandon the markup stripping. // Detection will have to work on the unstripped input. // if (openTags < 5 || openTags / 5 < badTags || (fInputLen < 100 && fRawLength > 600)) { int limit = fRawLength; if (limit > kBufSize) { limit = kBufSize; } for (srci = 0; srci < limit; srci++) { fInputBytes[srci] = fRawInput[srci]; } fInputLen = srci; } // // Tally up the byte occurence statistics. // These are available for use by the various detectors. // Arrays.fill(fByteStats, (short) 0); for (srci = 0; srci < fInputLen; srci++) { int val = fInputBytes[srci] & 0x00ff; fByteStats[val]++; } fC1Bytes = false; for (int i = 0x80; i <= 0x9F; i += 1) { if (fByteStats[i] != 0) { fC1Bytes = true; break; } } } }





© 2015 - 2024 Weber Informatics LLC | Privacy Policy