Skip to content

Commit

Permalink
JavaDoc Fixed
Browse files Browse the repository at this point in the history
  • Loading branch information
laugimethods committed Feb 1, 2018
1 parent 0e83d81 commit 05faf39
Show file tree
Hide file tree
Showing 12 changed files with 29 additions and 18 deletions.
1 change: 1 addition & 0 deletions .gitignore
Original file line number Diff line number Diff line change
Expand Up @@ -33,3 +33,4 @@ keys/
repository/io/nats/java-nats-streaming/0.1.0-SNAPSHOT/maven-metadata-local.xml

.DS_Store
dependency-reduced-pom.xml
8 changes: 5 additions & 3 deletions pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -139,7 +139,7 @@
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-javadoc-plugin</artifactId>
<version>2.10.4</version>
<version>3.0.0</version>
<executions>
<execution>
<id>attach-javadocs</id>
Expand All @@ -148,14 +148,16 @@
<goal>aggregate</goal>
</goals>
<configuration>
<failOnError>true</failOnError>
<failOnError>false</failOnError>
<show>public</show>
<nohelp>true</nohelp>
<header>nats-connector-spark ${project.version}</header>
<footer>nats-connector-spark ${project.version}</footer>
<doctitle>nats-connector-spark ${project.version}</doctitle>
<!-- <reportOutputDirectory>./target</reportOutputDirectory> -->
<destDir>target/apidocs/docs/${project.version}/api</destDir>
<outputDirectory>target/apidocs/docs/${project.version}/api</outputDirectory>
<!-- See https://github.com/lettuce-io/lettuce-core/issues/39 -->
<additionalparam>-Xdoclint:none</additionalparam>
</configuration>
</execution>
</executions>
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -16,6 +16,7 @@
public class IntegerNatsSubscriber extends NatsSubscriber {

/**
* @param natsUrl
* @param id
* @param subject
* @param count
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -16,10 +16,10 @@ public class NatsStreamingPublisher extends NatsPublisher {
protected final String clusterID, clientID;

/**
* @param id
* @param clusterID
* @param clientID
* @param natsUrl
* @param id
* @param subject
* @param count
*/
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -27,9 +27,13 @@ public class NatsStreamingSubscriber<V> extends NatsSubscriber {
private Collection<V> data;

/**
* @param natsUrl
* @param id
* @param subject
* @param count
* @param clusterName
* @param clientName
* @param data
* @param type
*/
public NatsStreamingSubscriber(String natsUrl, String id, String subject, String clusterName, String clientName, Collection<V> data, Class<V> type) {
super(natsUrl, id, subject, data.size());
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -13,6 +13,7 @@
public class StandardNatsSubscriber extends NatsSubscriber {

/**
* @param natsUrl
* @param id
* @param subject
* @param count
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -248,6 +248,8 @@ public static JavaPairDStream<String, String> getJavaPairDStream(final File temp

/**
* @param data
* @param subject
* @param url
* @return
*/
public static StandardNatsSubscriber getStandardNatsSubscriber(final List<Integer> data, String subject, String url) {
Expand All @@ -265,6 +267,10 @@ public static StandardNatsSubscriber getStandardNatsSubscriber(final List<Intege

/**
* @param data
* @param subject
* @param clusterName
* @param clientName
* @param url
* @return
*/
public static NatsStreamingSubscriber<Integer> getNatsStreamingSubscriber(final List<Integer> data, String subject, String clusterName, String clientName, String url) {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -69,13 +69,6 @@ protected NatsToSparkConnector(Class<V> type, StorageLevel storageLevel, String.
this.subjects = NatsSparkUtilities.transformIntoAList(subjects);
}

/**
* @param storageLevel
* @param subjects
* @param properties
* @param queue
* @param natsUrl
*/
protected NatsToSparkConnector(Class<V> type, StorageLevel storageLevel, Collection<String> subjects, Properties properties, String queue, String natsUrl) {
super(storageLevel);
this.type = type;
Expand Down Expand Up @@ -118,7 +111,7 @@ public T withNatsURL(String natsURL) {
}

/**
* @param dataExtractor, the Data Extractor to set
* @param dataDecoder, the Data Decoder to set
* @return the connector itself
*/
@SuppressWarnings("unchecked")
Expand Down Expand Up @@ -157,6 +150,7 @@ public static <V extends Object> StandardNatsToSparkConnectorImpl<V> receiveFrom
*
* @param type, the Class of Object to expect to receive from NATS
* @param storageLevel, defines the StorageLevel used by Spark
* @param clusterID, used by NATS Streaming
* @return a NATS Streaming to Spark Connector
*/
public static <V extends Object> NatsStreamingToSparkConnectorImpl<V> receiveFromNatsStreaming(Class<V> type, StorageLevel storageLevel, String clusterID) {
Expand Down Expand Up @@ -184,7 +178,8 @@ public void onStop() {
}

/** Create a socket connection and receive data until receiver is stopped
* @throws Exception **/
* @throws Exception
**/
protected abstract void receive() throws Exception;

protected void setQueue() {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -70,8 +70,7 @@ protected StandardNatsToKeyValueSparkConnectorImpl<V> storedAsKeyValue() {
/** Create a socket connection and receive data until receiver is stopped
* @throws IncompleteException
* @throws TimeoutException
* @throws IOException
* @throws Exception **/
* @throws IOException **/
protected void receive() throws IncompleteException, IOException, TimeoutException {

// Make connection and initialize streams
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -92,7 +92,7 @@ public static byte[] encodeData(Object obj) {
* @param type, the class of the object to decode
* @param bytes, the content that represent the object to decode
* @return the extracted object
* @throws UnsupportedOperationException, raised when the expected type of the object is not a Number or a String
* @throws UnsupportedOperationException , raised when the expected type of the object is not a Number or a String
* @see <a href="https://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html">Class ByteBuffer</a>
*/
@SuppressWarnings("unchecked")
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -118,7 +118,7 @@ public T withNatsURL(String natsURL) {
}

/**
* @param connectionTimeout, the connectionTimeout to set
* @param duration, the duration to set
* @return the connector itself
*/
@SuppressWarnings("unchecked")
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -66,6 +66,7 @@ public abstract class SparkToNatsConnectorPool<T> extends AbstractSparkToNatsCon

/**
* Create a pool of SparkToNatsConnector(s).
* @return a new SparkToStandardNatsConnectorPool
* @see <a href="http://spark.apache.org/docs/latest/streaming-programming-guide.html#design-patterns-for-using-foreachrdd">Design Patterns for using foreachRDD</a>
*/
public static SparkToStandardNatsConnectorPool newPool() {
Expand All @@ -76,6 +77,7 @@ public static SparkToStandardNatsConnectorPool newPool() {
/**
* Create a pool of SparkToNatsStreamingConnector(s).
* @param clusterID, the ID of the NATS Cluster
* @return a new SparkToNatsStreamingConnectorPool
* @see <a href="http://spark.apache.org/docs/latest/streaming-programming-guide.html#design-patterns-for-using-foreachrdd">Design Patterns for using foreachRDD</a>
*/
public static SparkToNatsStreamingConnectorPool newStreamingPool(String clusterID) {
Expand Down

0 comments on commit 05faf39

Please sign in to comment.