All Downloads are FREE. Search and download functionalities are using the official Maven repository.

io.anserini.index.generator.TweetGenerator Maven / Gradle / Ivy

/*
 * Anserini: A Lucene toolkit for reproducible information retrieval research
 *
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at
 *
 * http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */

package io.anserini.index.generator;

import com.twitter.twittertext.Extractor;
import com.twitter.twittertext.TwitterTextParseResults;
import com.twitter.twittertext.TwitterTextParser;
import io.anserini.collection.TweetCollection;
import io.anserini.index.IndexArgs;
import it.unimi.dsi.fastutil.longs.LongOpenHashSet;
import org.apache.logging.log4j.LogManager;
import org.apache.logging.log4j.Logger;
import org.apache.lucene.document.Document;
import org.apache.lucene.document.Field;
import org.apache.lucene.document.FieldType;
import org.apache.lucene.document.IntPoint;
import org.apache.lucene.document.LongPoint;
import org.apache.lucene.document.NumericDocValuesField;
import org.apache.lucene.document.StoredField;
import org.apache.lucene.document.StringField;
import org.apache.lucene.index.IndexOptions;
import org.apache.tools.bzip2.CBZip2InputStream;

import java.io.BufferedReader;
import java.io.File;
import java.io.FileInputStream;
import java.io.IOException;
import java.io.InputStreamReader;
import java.util.List;

/**
 * Converts a {@link TweetCollection.Document} into a Lucene {@link Document}, ready to be indexed.
 */
public class TweetGenerator implements LuceneDocumentGenerator {
  private static final Logger LOG = LogManager.getLogger(TweetGenerator.class);

  private IndexArgs args;

  private LongOpenHashSet deletes = null;

  public enum TweetField {
    ID_LONG("id_long"),
    SCREEN_NAME("screen_name"),
    EPOCH("epoch"),
    LANG("lang"),
    IN_REPLY_TO_STATUS_ID("in_reply_to_status_id"),
    IN_REPLY_TO_USER_ID("in_reply_to_user_id"),
    FOLLOWERS_COUNT("followers_count"),
    FRIENDS_COUNT("friends_count"),
    STATUSES_COUNT("statuses_count"),
    RETWEETED_STATUS_ID("retweeted_status_id"),
    RETWEETED_USER_ID("retweeted_user_id"),
    RETWEET_COUNT("retweet_count");

    public final String name;

    TweetField(String s) {
      name = s;
    }
  }

  public TweetGenerator(IndexArgs args) throws IOException {
    this.args = args;

    if (!args.tweetDeletedIdsFile.isEmpty()) {
      deletes = new LongOpenHashSet();
      File deletesFile = new File(args.tweetDeletedIdsFile);
      if (!deletesFile.exists()) {
        System.err.println("Error: " + deletesFile + " does not exist!");
        System.exit(-1);
      }
      LOG.info("Reading deletes from " + deletesFile);

      FileInputStream fin = new FileInputStream(deletesFile);
      byte[] ignoreBytes = new byte[2];
      fin.read(ignoreBytes); // "B", "Z" bytes from commandline tools
      BufferedReader br = new BufferedReader(new InputStreamReader(new CBZip2InputStream(fin)));

      String s;
      while ((s = br.readLine()) != null) {
        if (s.contains("\t")) {
          deletes.add(Long.parseLong(s.split("\t")[0]));
        } else {
          deletes.add(Long.parseLong(s));
        }
      }
      br.close();
      fin.close();
      LOG.info("Read " + deletes.size() + " tweetids from deletes file.");
    }
  }

  @Override
  public Document createDocument(TweetCollection.Document tweetDoc) throws GeneratorException {
    String id = tweetDoc.id();

    if (tweetDoc.contents().trim().isEmpty()) {
      throw new EmptyDocumentException();
    }

    final TwitterTextParseResults result = TwitterTextParser.parseTweet(tweetDoc.contents().trim());
    if (!result.isValid) {
      throw new InvalidDocumentException();
    }

    String text = tweetDoc.contents().trim().substring(result.validTextRange.start, result.validTextRange.end);

    if (!args.tweetKeepUrls) {
      final Extractor extractor = new Extractor();
      final List urls = extractor.extractURLs(text);
      for (String url : urls) {
        text = text.replaceAll(url, "");
      }
    }

    text = text.trim();
    if (text.isEmpty()) {
      throw new EmptyDocumentException();
    }

    // Skip deletes tweetids.
    if (deletes != null && deletes.contains(Long.parseLong(id))) {
      throw new SkippedDocumentException();
    }

    if (tweetDoc.getIdLong() > args.tweetMaxId) {
      throw new SkippedDocumentException();
    }

    if (!args.tweetKeepRetweets && tweetDoc.getRetweetedStatusId().isPresent()) {
      throw new SkippedDocumentException();
    }

    Document doc = new Document();
    doc.add(new StringField(IndexArgs.ID, id, Field.Store.YES));

    // We need this to break scoring ties.
    doc.add(new LongPoint(TweetField.ID_LONG.name, tweetDoc.getIdLong()));
    doc.add(new NumericDocValuesField(TweetField.ID_LONG.name, tweetDoc.getIdLong()));

    tweetDoc.getEpoch().ifPresent(epoch -> doc.add(new LongPoint(TweetField.EPOCH.name, epoch)));
    doc.add(new StringField(TweetField.SCREEN_NAME.name, tweetDoc.getScreenName(), Field.Store.NO));
    doc.add(new IntPoint(TweetField.FRIENDS_COUNT.name, tweetDoc.getFollowersCount()));
    doc.add(new IntPoint(TweetField.FOLLOWERS_COUNT.name, tweetDoc.getFriendsCount()));
    doc.add(new IntPoint(TweetField.STATUSES_COUNT.name, tweetDoc.getStatusesCount()));

    tweetDoc.getInReplyToStatusId().ifPresent(rid -> {
      doc.add(new LongPoint(TweetField.IN_REPLY_TO_STATUS_ID.name, rid));
      tweetDoc.getInReplyToUserId().ifPresent(ruid ->
        doc.add(new LongPoint(TweetField.IN_REPLY_TO_USER_ID.name, ruid)));
    });

    tweetDoc.getRetweetedStatusId().ifPresent(rid -> {
      doc.add(new LongPoint(TweetField.RETWEETED_STATUS_ID.name, rid));
      tweetDoc.getRetweetedUserId().ifPresent(ruid ->
        doc.add(new LongPoint(TweetField.RETWEETED_USER_ID.name, ruid)));
      tweetDoc.getRetweetCount().ifPresent(rc ->
        doc.add(new LongPoint(TweetField.RETWEET_COUNT.name, rc)));
    });

    tweetDoc.getLang().ifPresent(lang -> doc.add(new StringField(TweetField.LANG.name, lang, Field.Store.NO)));

    if (args.storeRaw) { // store the raw json string as one single field
      doc.add(new StoredField(IndexArgs.RAW, tweetDoc.getJsonString()));
    }

    FieldType fieldType = new FieldType();

    fieldType.setStored(args.storeContents);

    // Are we storing document vectors?
    if (args.storeDocvectors) {
      fieldType.setStoreTermVectors(true);
      fieldType.setStoreTermVectorPositions(true);
    }

    // Are we building a "positional" or "count" index?
    if (args.storePositions) {
      fieldType.setIndexOptions(IndexOptions.DOCS_AND_FREQS_AND_POSITIONS);
    } else {
      fieldType.setIndexOptions(IndexOptions.DOCS_AND_FREQS);
    }

    doc.add(new Field(IndexArgs.CONTENTS, text, fieldType));

    return doc;
  }
}




© 2015 - 2025 Weber Informatics LLC | Privacy Policy