Skip to content

Commit

Permalink
Remove all javadoc warnings
Browse files Browse the repository at this point in the history
  • Loading branch information
mujtabachohan committed Feb 21, 2013
1 parent 3ddcf0f commit a7083ca
Show file tree
Hide file tree
Showing 35 changed files with 49 additions and 73 deletions.
9 changes: 2 additions & 7 deletions build.txt
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,7 @@ Phoenix uses Maven (3.X) to build all its necessary resources.
and optionally, to just skip all the tests and build the jars:
$ mvn package -DskipTests

3.
Note: javadocs are generated in target/apidocs

## Importing into eclipse
=========================
Expand All @@ -27,13 +27,8 @@ Use the m2e eclipse plugin and do Import->Maven Project and just pick the root '
1. All tests
$ mvn clean test

## Javadocs
===========
Javadoc are generated in /target/site/apidocs
$ mvn site

## Findbugs
===========
Findbugs report is generated in /target/site
$ mvn site


26 changes: 13 additions & 13 deletions pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -268,19 +268,19 @@
</executions>
</plugin>
<plugin>
<artifactId>maven-javadoc-plugin</artifactId>
<version>2.6.1</version>
<configuration>
<docfilessubdirs>true</docfilessubdirs>
</configuration>
<executions>
<execution>
<phase>site</phase>
<goals>
<goal>javadoc</goal>
</goals>
</execution>
</executions>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-javadoc-plugin</artifactId>
<configuration>
<quiet>true</quiet>
</configuration>
<executions>
<execution>
<id>attach-javadocs</id>
<goals>
<goal>jar</goal>
</goals>
</execution>
</executions>
</plugin>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -87,7 +87,7 @@ public ConcurrentHashMap<ImmutableBytesPtr,PTable> getMetaDataCache() {
* used in which case a global tenant cache is returned.
* @param config the HBase configuration
* @param tenantId the tenant ID or null if not applicable.
* @return
* @return TenantCache
*/
public static TenantCache getTenantCache(Configuration config, ImmutableBytesWritable tenantId) {
GlobalCache globalCache = GlobalCache.getInstance(config);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -67,7 +67,6 @@ public interface ColumnProjector {
* Get the value of the column, coercing it if necessary to the specified type
* @param tuple the row containing the column
* @param type the type to which to coerce the binary value
* @param ptr TODO
* @param ptr used to retrieve the value
* @return the object representation of the column value.
* @throws SQLException
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -301,8 +301,6 @@ private Expression wrapGroupByExpression(Expression expression) {
/**
* Add a Function expression to the expression manager.
* Derived classes may use this as a hook to trap all function additions.
* @param node a function expression node
* @param children the child expression arguments to the function expression node.
* @return a Function expression
* @throws SQLException if the arguments are invalid for the function.
*/
Expand All @@ -311,6 +309,10 @@ protected Expression addFunction(FunctionExpression func) {
}

@Override
/**
* @param node a function expression node
* @param children the child expression arguments to the function expression node.
*/
public Expression visitLeave(FunctionParseNode node, List<Expression> children) throws SQLException {
children = node.validate(children, context);
FunctionExpression func = node.create(children, context);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -82,7 +82,6 @@ public static ColumnResolver getResolver(DropColumnStatement statement, PhoenixC
* Iterate through the nodes in the FROM clause to build a column resolver used to
* lookup a column given the name and alias.
* @param statement the select statement
* @param schema the schema
* @return the column resolver
* @throws SQLException
* @throws SQLFeatureNotSupportedException if unsupported constructs appear in the FROM
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -75,7 +75,7 @@ public List<OrderingColumn> getOrderingColumns() {
* @param context the query context for tracking various states
* associated with the given select statement
* @param limit
* @param groupByExpressions the list of columns in the GROUP BY clause
* @param groupBy the list of columns in the GROUP BY clause
* @return the list of columns in the ORDER BY clause
* @throws SQLException
*/
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -73,7 +73,7 @@ private ProjectionCompiler() {
* @param statement SQL statement being compiled
* @param context query context kept between compilation of different query clauses
* @param limit maximum number of rows to scan during query execution or null if unbounded
* @param groupByExpressions list of GROUP BY expressions or the empty list if no GROUP BY
* @param groupBy list of GROUP BY expressions or the empty list if no GROUP BY
* @return projector used to access row values during scan
* @throws SQLException
* @throws SQLFeatureNotSupportedException if an unsupported construct is encountered.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -81,7 +81,6 @@ public QueryCompiler(PhoenixConnection connection, int maxRows, Scan scan, PColu
* Builds an executable query plan from a parsed SQL statement
* @param statement parsed SQL statement
* @param binds values of bind variables
* @param scan TODO
* @return executable query plan
* @throws SQLException if mismatched types are found, bind value do not match binds,
* or invalid function arguments are encountered.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -54,7 +54,6 @@ public class RowProjector {
/**
* Construct RowProjector based on a list of ColumnProjectors.
* @param columnProjectors ordered list of ColumnProjectors corresponding to projected columns in SELECT clause
* @param rowCountExpression Expression used to calculate the row count from the result returned by the server-side
* aggregating coprocessor. Only required in the case of an aggregate query with a limit clause and otherwise may
* be null.
*/
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -48,7 +48,6 @@ public ExplainPlan getExplainPlan() throws SQLException {

/**
* Returns the ParameterMetaData for the statement
* @return
*/
ParameterMetaData getParameterMetaData();

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -129,38 +129,38 @@ public void write(DataOutput output) throws IOException {
* @param tableName
* @param tableTimestamp
* @param clientTimestamp
* @return
* @return MetaDataMutationResult
* @throws IOException
*/
MetaDataMutationResult getTable(byte[] schemaName, byte[] tableName, long tableTimestamp, long clientTimestamp) throws IOException;
/**
* Create a new Phoenix table
* @param tableMetadata
* @return
* @return MetaDataMutationResult
* @throws IOException
*/
MetaDataMutationResult createTable(List<Mutation> tableMetadata) throws IOException;
/**
* Drop an existing Phoenix table
* @param tableMetadata
* @param isView TODO
* @return
* @return MetaDataMutationResult
* @throws IOException
*/
MetaDataMutationResult dropTable(List<Mutation> tableMetadata, boolean isView) throws IOException;

/**
* Add a column to an existing Phoenix table
* @param tableMetadata
* @return
* @return MetaDataMutationResult
* @throws IOException
*/
MetaDataMutationResult addColumn(List<Mutation> tableMetadata) throws IOException;

/**
* Drop a column from an existing Phoenix table
* @param tableMetadata
* @return
* @return MetaDataMutationResult
* @throws IOException
*/
MetaDataMutationResult dropColumn(List<Mutation> tableMetadata) throws IOException;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -55,8 +55,7 @@ public class AggregateRowCounter implements RowCounter {
/**
* Construct an AggregateRowCounter that is used to short circuit scans with a rownum limit by
* calculating the row count from aggregated query results.
* @param valueSchema the schema of the row result value.
* @param rowCountExpression expression that calculates the row count from a given row result
* @param aggregators aggregators
*/
public AggregateRowCounter(Aggregators aggregators) {
this.schema = aggregators.getValueSchema();
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -95,10 +95,10 @@ private ConnectionQueryServices getConnectionQueryServices(ConnectionQueryServic
return childServices;
}

/**
* Sets up an id used to do round robin queue processing on the server
* @param scan
*/
// /**
// * Sets up an id used to do round robin queue processing on the server
// * @param scan
// */
// private void setProducer(Scan scan) {
// byte[] producer = Bytes.toBytes(UUID.randomUUID().toString());
// scan.setAttribute(HBaseServer.CALL_QUEUE_PRODUCER_ATTRIB_NAME, producer);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -46,8 +46,6 @@ public interface Aggregator extends Expression {
* Incrementally aggregate the value with the current row
* @param tuple the result containing all the key values of the row
* @param ptr the bytes pointer to the underlying result
* @return true if the child expression of the aggregator was
* able to be evaluated and false otherwise.
*/
public void aggregate(Tuple tuple, ImmutableBytesWritable ptr);

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -70,7 +70,7 @@ public void aggregate(Aggregator[] aggregators, Tuple result) {

/**
* Serialize an Aggregator into a byte array
* @param aggregator the aggregator to serialize
* @param aggFuncs list of aggregator to serialize
* @return serialized byte array respresentation of aggregator
*/
public static byte[] serialize(List<SingleAggregateFunction> aggFuncs, int minNullableIndex) {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -101,7 +101,7 @@ public KeyFormationDirective getKeyFormationDirective() {
* Zero-based index of child expression that determines the potential column used to form the row key during optimization.
* For example, SUBSTR(prefix,1,3), the first child expression (prefix) would determine the row key column used to
* help form the scan start/stop key
* @return
* @return int
*/
public int getKeyFormationTraversalIndex() {
return 0;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -49,10 +49,10 @@
* REGEXP_REPLACE(<source_char>,<pattern>,<replace_string>)
* source_char is the string in which we want to perform string replacement. pattern is a
* Java compatible regular expression string, and we replace all the matching part with
* replace_string. The first 2 arguments are required and are {@link com.salesforce.phoenix.schema.PDataType.VARCHAR},
* replace_string. The first 2 arguments are required and are {@link com.salesforce.phoenix.schema.PDataType#VARCHAR},
* the replace_string is default to empty string.
*
* The function returns a {@link com.salesforce.phoenix.schema.PDataType.VARCHAR}
* The function returns a {@link com.salesforce.phoenix.schema.PDataType#VARCHAR}
*
* @author zhuang
* @since 0.1
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -50,7 +50,7 @@
* an even increment. Usage:
* ROUND(<date/time col ref>,<'day'|'hour'|'minute'|'second'|'millisecond'>,<optional integer multiplier>)
* The integer multiplier is optional and is used to do rollups to a partial time unit (i.e. 10 minute rollup)
* The function returns a {@link com.salesforce.phoenix.schema.PDataType.DATE}
* The function returns a {@link com.salesforce.phoenix.schema.PDataType#DATE}
*
* @author jtaylor
* @since 0.1
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -42,7 +42,7 @@
* an even increment. Usage:
* TRUNC(<date/time col ref>,<'day'|'hour'|'minute'|'second'|'millisecond'>,[<optional integer multiplier>])
* The integer multiplier is optional and is used to do rollups to a partial time unit (i.e. 10 minute rollup)
* The function returns a {@link com.salesforce.phoenix.schema.PDataType.DATE}
* The function returns a {@link com.salesforce.phoenix.schema.PDataType#DATE}
*
* @author jtaylor
* @since 0.1
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -41,7 +41,7 @@
public interface AggregatingResultIterator extends ResultIterator {
/**
* Provides a means of re-aggregating a result row. For
* scanners that need to look ahead (i.e. {@link phoenix.execute.iterate.OrderedAggregatingResultIterator.OrderByResultScanner}
* scanners that need to look ahead (i.e. {@link com.salesforce.phoenix.iterate.OrderedAggregatingResultIterator}
* @param result the row to re-aggregate
*/
void aggregate(Tuple result);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -43,7 +43,7 @@
* Post aggregation filter for HAVING clause. Due to the way we cache aggregation values
* we cannot have a look ahead for this Iterator, because the expressions in the SELECT
* clause would return values for the peeked row instead of the current row. If we only
* use the Result argument in {@link com.salesforce.phoenix.expression.Expression#getValue(Result)}
* use the Result argument in {@link com.salesforce.phoenix.expression.Expression}
* instead of our cached value in Aggregators, we could have a look ahead.
*
* @author jtaylor
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -106,8 +106,6 @@ public boolean apply(HRegionInfo region) {
/**
* Splits the given scan's key range so that each split can be queried in parallel
*
* @param config configuration object that holds concurrency settings for
* the {@link ParallelIterators}.
* @param scan the scan to parallelize
* @param allTableRegions all online regions for the table to be scanned
* @return the key ranges that should be scanned in parallel
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -275,7 +275,7 @@ public Statement createStatement() throws SQLException {
/**
* Back-door way to inject processing into walking through a result set
* @param statementFactory
* @return
* @return PhoenixStatement
* @throws SQLException
*/
public PhoenixStatement createStatement(PhoenixStatementFactory statementFactory) throws SQLException {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -76,9 +76,8 @@ public class HashCacheClient {
* Construct client used to create a serialized cached snapshot of a table and send it to each region server
* for caching during hash join processing.
* @param services the global services
* @param memoryManager the per request memory manager
* @param loopTable the table being iterated over (as opposed to cached) during join processing
* @param joinKeyPrefix bytes prefixing key of every LHS and RHS row
* @param iterateOverTableName table name
* @param tenantId the tenantId or null if not applicable
*/
public HashCacheClient(ConnectionQueryServices services, byte[] iterateOverTableName, byte[] tenantId) {
this.services = services;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -216,7 +216,6 @@ public List<Expression> validate(List<Expression> children, StatementContext con

/**
* Entry point for parser to instantiate compiled representation of built-in function
* @param node Parse model node representing a built-in function reference
* @param children Compiled expressions for child nodes
* @param context Query context for accessing state shared across the processing of multiple clauses
* @return compiled representation of built-in function
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -45,7 +45,6 @@ public class RHSLiteralStatementRewriter extends ParseNodeRewriter {
/**
* Rewrite the select statement by filtering out expression nodes from the WHERE clause
* @param statement the select statement from which to filter.
* @param removeNodes expression nodes to filter out of WHERE clause.
* @return new select statement
* @throws SQLException
*/
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -47,7 +47,7 @@ public interface ConnectionQueryServices extends QueryServices, MetaDataMutated
/**
* Get (and create if necessary) a child QueryService for a given tenantId.
* The QueryService will be cached for the lifetime of the parent QueryService
* @param childId the organization ID
* @param tenantId the organization ID
* @return the child QueryService
*/
public ConnectionQueryServices getChildQueryServices(ImmutableBytesWritable tenantId);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -29,7 +29,7 @@

/**
*
* Abstract class for columns of type {@link com.salesforce.phoenix.schema.PDataType.STRING}
* Abstract class for columns of type {@link com.salesforce.phoenix.schema.PDataType#VARCHAR}
*
* @author jtaylor
* @since 0.1
Expand Down
4 changes: 2 additions & 2 deletions src/main/java/com/salesforce/phoenix/schema/PTable.java
Original file line number Diff line number Diff line change
Expand Up @@ -106,7 +106,7 @@ public interface PTable extends Writable {

/**
* Creates a new row at the specified timestamp using the key
* for the PK values (from {@link #newKey(ImmutableBytesWritable, byte[]...)}
* for the PK values (from {@link #newKey(ImmutableBytesWritable, byte[][])}
* and the optional key values specified using values.
* @param ts the timestamp that the key value will have when committed
* @param key the row key of the key value
Expand All @@ -119,7 +119,7 @@ public interface PTable extends Writable {
PRow newRow(long ts, ImmutableBytesWritable key, byte[]... values);

/**
* Creates a new row for the PK values (from {@link #newKey(ImmutableBytesWritable, byte[]...)}
* Creates a new row for the PK values (from {@link #newKey(ImmutableBytesWritable, byte[][])}
* and the optional key values specified using values. The timestamp of the key value
* will be set by the HBase server.
* @param key the row key of the key value
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -174,7 +174,6 @@ public int getOffset(byte[] keyBuffer, int keyOffset) {

/**
* Calculate the length of the PK column value
* @param maxLength the fixed length of the PK column value or null if variable length
* @param keyBuffer the byte array of the row key
* @param keyOffset the offset in the byte array of where the key begins
* @param keyLength the length of the entire row key
Expand Down
2 changes: 1 addition & 1 deletion src/main/java/com/salesforce/phoenix/util/CSVLoader.java
Original file line number Diff line number Diff line change
Expand Up @@ -84,7 +84,7 @@ public void upsert(String fileName) throws Exception {
* column value to correct type before upsert. Note: Column Names are
* expected as first line of CSV file.
*
* @param fileName
* @param reader CSVReader instance
* @throws Exception
*/
public void upsert(CSVReader reader) throws Exception {
Expand Down
3 changes: 1 addition & 2 deletions src/main/java/com/salesforce/phoenix/util/KeyValueUtil.java
Original file line number Diff line number Diff line change
Expand Up @@ -92,9 +92,8 @@ public static KeyValue newKeyValue(Result r, byte[] cf, byte[] cq, long ts, byte

/**
* Binary search for latest column value without allocating memory in the process
* @param r
* @param kvs
* @param searchTerm
* @return
*/
public static KeyValue getColumnLatest(List<KeyValue>kvs, KeyValue searchTerm) {
if (kvs.size() == 0) {
Expand Down
Loading

0 comments on commit a7083ca

Please sign in to comment.