net.sourceforge.pmd.cpd.Tokens Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of pmd-core Show documentation
Show all versions of pmd-core Show documentation
PMD is an extensible multilanguage static code analyzer. It finds common programming flaws like unused variables,
empty catch blocks, unnecessary object creation, and so forth. It's mainly concerned with Java and
Apex, but supports 16 other languages. It comes with 400+ built-in rules. It can be
extended with custom rules. It uses JavaCC and Antlr to parse source files into abstract syntax trees
(AST) and runs rules against them to find violations. Rules can be written in Java or using a XPath query.
Currently, PMD supports Java, JavaScript, Salesforce.com Apex and Visualforce,
Kotlin, Swift, Modelica, PLSQL, Apache Velocity, JSP, WSDL, Maven POM, HTML, XML and XSL.
Scala is supported, but there are currently no Scala rules available.
Additionally, it includes CPD, the copy-paste-detector. CPD finds duplicated code in
Coco, C/C++, C#, Dart, Fortran, Gherkin, Go, Groovy, HTML, Java, JavaScript, JSP, Julia, Kotlin,
Lua, Matlab, Modelica, Objective-C, Perl, PHP, PLSQL, Python, Ruby, Salesforce.com Apex and
Visualforce, Scala, Swift, T-SQL, Typescript, Apache Velocity, WSDL, XML and XSL.
The newest version!
/**
* BSD-style license; for more info see http://pmd.sourceforge.net/license.html
*/
package net.sourceforge.pmd.cpd;
import java.util.ArrayList;
import java.util.HashMap;
import java.util.List;
import java.util.Map;
import java.util.Map.Entry;
import org.checkerframework.checker.nullness.qual.NonNull;
import org.checkerframework.checker.nullness.qual.Nullable;
import net.sourceforge.pmd.lang.ast.LexException;
import net.sourceforge.pmd.lang.document.FileId;
import net.sourceforge.pmd.lang.document.TextDocument;
/**
* Global token collector for CPD. This is populated by lexing all files,
* after which the match algorithm proceeds.
*/
public class Tokens {
// This stores all the token entries recorded during the run.
private final List tokens = new ArrayList<>();
private final Map images = new HashMap<>();
// the first ID is 1, 0 is the ID of the EOF token.
private int curImageId = 1;
/**
* Create a new instance.
*
* @apiNote Internal API
*/
Tokens() {
// constructor is package private
}
private void add(TokenEntry tokenEntry) {
this.tokens.add(tokenEntry);
}
private void addEof(FileId filePathId, int line, int column) {
add(new TokenEntry(filePathId, line, column));
}
private void setImage(TokenEntry entry, String newImage) {
int i = getImageId(newImage);
entry.setImageIdentifier(i);
}
private int getImageId(String newImage) {
return images.computeIfAbsent(newImage, k -> curImageId++);
}
String imageFromId(int i) {
return images.entrySet().stream().filter(it -> it.getValue() == i).findFirst().map(Entry::getKey).orElse(null);
}
private TokenEntry peekLastToken() {
return tokens.isEmpty() ? null : getToken(size() - 1);
}
private TokenEntry getToken(int index) {
return tokens.get(index);
}
public int size() {
return tokens.size();
}
TokenEntry getEndToken(TokenEntry mark, Match match) {
return getToken(mark.getIndex() + match.getTokenCount() - 1);
}
public List getTokens() {
return tokens;
}
TokenEntry addToken(String image, FileId fileName, int startLine, int startCol, int endLine, int endCol) {
TokenEntry newToken = new TokenEntry(getImageId(image), fileName, startLine, startCol, endLine, endCol, tokens.size());
add(newToken);
return newToken;
}
State savePoint() {
return new State(this);
}
/**
* Creates a token factory to process the given file with
* {@link CpdLexer#tokenize(TextDocument, TokenFactory)}.
* Tokens are accumulated in the {@link Tokens} parameter.
*
* @param file Document for the file to process
* @param tokens Token sink
*
* @return A new token factory
*/
static TokenFactory factoryForFile(TextDocument file, Tokens tokens) {
return new TokenFactory() {
final FileId fileId = file.getFileId();
final int firstToken = tokens.size();
@Override
public void recordToken(@NonNull String image, int startLine, int startCol, int endLine, int endCol) {
tokens.addToken(image, fileId, startLine, startCol, endLine, endCol);
}
@Override
public void setImage(TokenEntry entry, @NonNull String newImage) {
tokens.setImage(entry, newImage);
}
@Override
public LexException makeLexException(int line, int column, String message, @Nullable Throwable cause) {
return new LexException(line, column, fileId, message, cause);
}
@Override
public @Nullable TokenEntry peekLastToken() {
if (tokens.size() <= firstToken) {
return null; // no token has been added yet in this file
}
return tokens.peekLastToken();
}
@Override
public void close() {
TokenEntry tok = peekLastToken();
if (tok == null) {
tokens.addEof(fileId, 1, 1);
} else {
tokens.addEof(fileId, tok.getEndLine(), tok.getEndColumn());
}
}
};
}
/**
* Helper class to preserve and restore the current state of the token
* entries.
*/
static final class State {
private final int tokenCount;
private final int curImageId;
State(Tokens tokens) {
this.tokenCount = tokens.tokens.size();
this.curImageId = tokens.curImageId;
}
public void restore(Tokens tokens) {
tokens.images.entrySet().removeIf(e -> e.getValue() >= curImageId);
tokens.curImageId = this.curImageId;
final List entries = tokens.getTokens();
entries.subList(tokenCount, entries.size()).clear();
}
}
}