Datasets:
The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
Error code: DatasetGenerationCastError Exception: DatasetGenerationCastError Message: An error occurred while generating the dataset All the data files must have the same columns, but at some point there are 10 new columns ({'Parameters', 'DevelopmentNotes', 'Version', 'Todo', 'Exception', 'Warnings', 'Links', 'Dependecies', 'Summary', 'Usage'}) and 15 missing columns ({'usage', 'Incomplete', 'exception', 'summary', 'Commentedcode', 'todo', 'Ownership', 'Pointer', 'rational', 'License', 'deprecation', 'Autogenerated', 'formatter', 'directive', 'Warning'}). This happened while the csv dataset builder was generating data using hf://datasets/poojaruhal/Code-comment-classification/python_0_raw.csv (at revision 3d2bbff4d30d5c41d2cbf5b1d55fbc8d10cfdbaa) Please either edit the data files to have matching columns, or separate them into different configurations (see docs at https://hf.co/docs/hub/datasets-manual-configuration#multiple-configurations) Traceback: Traceback (most recent call last): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 2011, in _prepare_split_single writer.write_table(table) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 585, in write_table pa_table = table_cast(pa_table, self._schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2302, in table_cast return cast_table_to_schema(table, schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2256, in cast_table_to_schema raise CastError( datasets.table.CastError: Couldn't cast stratum: int64 class: string comment: string Summary: string Usage: string Parameters: string Expand: string Version: string DevelopmentNotes: string Todo: string Exception: string Links: string Noise: string Warnings: string Recommendation: string Dependecies: string Precondition: string CodingGuidelines: string Extension: string Subclassexplnation: string Observation: string -- schema metadata -- pandas: '{"index_columns": [{"kind": "range", "name": null, "start": 0, "' + 2742 to {'stratum': Value(dtype='int64', id=None), 'class': Value(dtype='string', id=None), 'comment': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None), 'Expand': Value(dtype='string', id=None), 'rational': Value(dtype='string', id=None), 'deprecation': Value(dtype='string', id=None), 'usage': Value(dtype='string', id=None), 'exception': Value(dtype='string', id=None), 'todo': Value(dtype='string', id=None), 'Incomplete': Value(dtype='string', id=None), 'Commentedcode': Value(dtype='string', id=None), 'directive': Value(dtype='string', id=None), 'formatter': Value(dtype='string', id=None), 'License': Value(dtype='float64', id=None), 'Ownership': Value(dtype='string', id=None), 'Pointer': Value(dtype='string', id=None), 'Autogenerated': Value(dtype='string', id=None), 'Noise': Value(dtype='string', id=None), 'Warning': Value(dtype='string', id=None), 'Recommendation': Value(dtype='string', id=None), 'Precondition': Value(dtype='string', id=None), 'CodingGuidelines': Value(dtype='float64', id=None), 'Extension': Value(dtype='float64', id=None), 'Subclassexplnation': Value(dtype='string', id=None), 'Observation': Value(dtype='string', id=None)} because column names don't match During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1321, in compute_config_parquet_and_info_response parquet_operations = convert_to_parquet(builder) File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 935, in convert_to_parquet builder.download_and_prepare( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1027, in download_and_prepare self._download_and_prepare( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1122, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1882, in _prepare_split for job_id, done, content in self._prepare_split_single( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 2013, in _prepare_split_single raise DatasetGenerationCastError.from_cast_error( datasets.exceptions.DatasetGenerationCastError: An error occurred while generating the dataset All the data files must have the same columns, but at some point there are 10 new columns ({'Parameters', 'DevelopmentNotes', 'Version', 'Todo', 'Exception', 'Warnings', 'Links', 'Dependecies', 'Summary', 'Usage'}) and 15 missing columns ({'usage', 'Incomplete', 'exception', 'summary', 'Commentedcode', 'todo', 'Ownership', 'Pointer', 'rational', 'License', 'deprecation', 'Autogenerated', 'formatter', 'directive', 'Warning'}). This happened while the csv dataset builder was generating data using hf://datasets/poojaruhal/Code-comment-classification/python_0_raw.csv (at revision 3d2bbff4d30d5c41d2cbf5b1d55fbc8d10cfdbaa) Please either edit the data files to have matching columns, or separate them into different configurations (see docs at https://hf.co/docs/hub/datasets-manual-configuration#multiple-configurations)
Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
stratum
int64 | class
string | comment
string | summary
string | Expand
string | rational
string | deprecation
null | usage
null | exception
null | todo
null | Incomplete
null | Commentedcode
null | directive
null | formatter
null | License
null | Ownership
null | Pointer
null | Autogenerated
null | Noise
null | Warning
null | Recommendation
null | Precondition
null | CodingGuidelines
null | Extension
null | Subclassexplnation
null | Observation
null |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
1 | Abfss.java | * Azure Blob File System implementation of AbstractFileSystem.
* This impl delegates to the old FileSystem | Azure Blob File System implementation of AbstractFileSystem. | This impl delegates to the old FileSystem | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | AbstractContractGetFileStatusTest.java | * Test getFileStatus and related listing operations.
| the tree parameters. Kept small to avoid killing object store test| runs too much.|
* Accept everything.
| the tree parameters. Kept small to avoid killing object store test| runs too much.|
* Accept nothing.
| the tree parameters. Kept small to avoid killing object store test| runs too much.|
* Path filter which only expects paths whose final name element
* equals the {@code match} field.
| the tree parameters. Kept small to avoid killing object store test| runs too much.|
* A filesystem filter which exposes the protected method
* {@link #listLocatedStatus(Path, PathFilter)}. | Test getFileStatus and related listing operations. | | the tree parameters. Kept small to avoid killing object store test| runs too much.|
* Accept everything.
| the tree parameters. Kept small to avoid killing object store test| runs too much.|
* Accept nothing.
| the tree parameters. Kept small to avoid killing object store test| runs too much.|
* Path filter which only expects paths whose final name element
* equals the {@code match} field.
| the tree parameters. Kept small to avoid killing object store test| runs too much.|
* A filesystem filter which exposes the protected method
* {@link #listLocatedStatus(Path, PathFilter)}. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | AbstractFSContract.java | * Class representing a filesystem contract that a filesystem
* implementation is expected implement.
*
* Part of this contract class is to allow FS implementations to
* provide specific opt outs and limits, so that tests can be
* skip unsupported features (e.g. case sensitivity tests),
* dangerous operations (e.g. trying to delete the root directory),
* and limit filesize and other numeric variables for scale tests | Class representing a filesystem contract that a filesystem * implementation is expected implement. | null | Part of this contract class is to allow FS implementations to * provide specific opt outs and limits, so that tests can be * skip unsupported features (e.g. case sensitivity tests), * dangerous operations (e.g. trying to delete the root directory), * and limit filesize and other numeric variables for scale tests | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | AbstractS3ACommitterFactory.java | * Dynamically create the output committer based on subclass type and settings. | null | Dynamically create the output committer based on subclass type and settings. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | AbstractTFLaunchCommandTestHelper.java | * This class is an abstract base class for testing Tensorboard and TensorFlow
* launch commands. | This class is an abstract base class for testing Tensorboard and TensorFlow and launch commands. | null | This class is an abstract base class for testing Tensorboard and TensorFlow and launch commands. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | ApplicationConstants.java | * This is the API for the applications comprising of constants that YARN sets
* up for the applications and the containers.
*
* TODO: Investigate the semantics and security of each cross-boundary refs.
|
* The type of launch for the container.
|
* Environment for Applications.
*
* Some of the environment variables for applications are <em>final</em>
* i.e. they cannot be modified by the applications. | * This is the API for the applications comprising of constants that YARN sets * up for the applications and the containers. | The type of launch for the container. | * Environment for Applications. * * Some of the environment variables for applications are <em>final</em> * i.e. they cannot be modified by the applications. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | ApplicationStateData.java | * Contains all the state data that needs to be stored persistently * for an Application | Contains all the state data that needs to be stored persistently * for an Application | null | needs to be stored persistently * for an Application | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | AutoInputFormat.java | * An {@link InputFormat} that tries to deduce the types of the input files
* automatically. It can currently handle text and sequence files. | An {@link InputFormat} that tries to deduce the types of the input files * automatically. | It can currently handle text and sequence files. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | BalancingPolicy.java | * Balancing policy.
* Since a datanode may contain multiple block pools,
* {@link Pool} implies {@link Node}
* but NOT the other way around
|
* Cluster is balanced if each node is balanced.
|
* Cluster is balanced if each pool in each node is balanced. | * Balancing policy. | Since a datanode may contain multiple block pools, * {@link Pool} implies {@link Node} * but NOT the other way around | | Cluster is balanced if each node is balanced. | * Cluster is balanced if each pool in each node is balanced. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | BaseRouterWebServicesTest.java | * Base class for all the RouterRMAdminService test cases. It provides utility
* methods that can be used by the concrete test case classes.
* | * Base class for all the RouterRMAdminService test cases. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | BatchedRequests.java | * A grouping of Scheduling Requests which are sent to the PlacementAlgorithm
* to place as a batch. The placement algorithm tends to give more optimal
* placements if more requests are batched together.
| PlacementAlgorithmOutput attempt - the number of times the requests in this|
* Iterator Type. | A grouping of Scheduling Requests which are sent to the PlacementAlgorithm
* to place as a batch
Iterator Type. | The placement algorithm tends to give more optimal
* placements if more requests are batched together. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | BlockPlacementStatusWithNodeGroup.java | * An implementation of @see BlockPlacementStatus for
* @see BlockPlacementPolicyWithNodeGroup | An implementation of @see BlockPlacementStatus | null | for * @see BlockPlacementPolicyWithNodeGroup | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | BlocksMap.java | * This class maintains the map from a block to its metadata.
* block's metadata currently includes blockCollection it belongs to and
* the datanodes that store the block. | This class maintains the map from a block to its metadata. | block's metadata currently includes blockCollection it belongs to and
* the datanodes that store the block. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | BlockUtils.java | * Utils functions to help block functions. | Utils functions to help block functions | null | Utils functions to help block functions | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | ByteArrayEncodingState.java | * A utility class that maintains encoding state during an encode call using
* byte array inputs. | A utility class that maintains encoding state during an encode call using * byte array inputs. | A utility class that maintains encoding state during an encode call using * byte array inputs. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | CapacitySchedulerPlanFollower.java | * This class implements a {@link PlanFollower}. This is invoked on a timer, and
* it is in charge to publish the state of the {@link Plan}s to the underlying
* {@link CapacityScheduler}. This implementation does so, by
* adding/removing/resizing leaf queues in the scheduler, thus affecting the
* dynamic behavior of the scheduler in a way that is consistent with the
* content of the plan. It also updates the plan's view on how much resources
* are available in the cluster.
* * This implementation of PlanFollower is relatively stateless, and it can
* synchronize schedulers and Plans that have arbitrary changes (performing set
* differences among existing queues). This makes it resilient to frequency of
* synchronization, and RM restart issues (no "catch up" is necessary). | This class implements a {@link PlanFollower}. This is invoked on a timer, and
* it is in charge to publish the state of the {@link Plan}s to the underlying
* {@link CapacityScheduler}. | This implementation does so, by
* adding/removing/resizing leaf queues in the scheduler, thus affecting the
* dynamic behavior of the scheduler in a way that is consistent with the
* content of the plan. It also updates the plan's view on how much resources
* are available in the cluster. | This implementation of PlanFollower is relatively stateless, and it can
* synchronize schedulers and Plans that have arbitrary changes (performing set
* differences among existing queues). This makes it resilient to frequency of
* synchronization, and RM restart issues (no "catch up" is necessary). | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | Classpath.java | * Command-line utility for getting the full classpath needed to launch a Hadoop
* client application. If the hadoop script is called with "classpath" as the
* command, then it simply prints the classpath and exits immediately without
* launching a JVM. The output likely will include wildcards in the classpath.
* If there are arguments passed to the classpath command, then this class gets
* called. With the --glob argument, it prints the full classpath with wildcards
* expanded. This is useful in situations where wildcard syntax isn't usable.
* With the --jar argument, it writes the classpath as a manifest in a jar file.
* This is useful in environments with short limitations on the maximum command
* line length, where it may not be possible to specify the full classpath in a
* command. For example, the maximum command line length on Windows is 8191
* characters. | Command-line utility for getting the full classpath needed to launch a Hadoop * client application. | If the hadoop script is called with "classpath" as the * command, then it simply prints the classpath and exits immediately without * launching a JVM. The output likely will include wildcards in the classpath. * If there are arguments passed to the classpath command, then this class gets * called. With the --glob argument, it prints the full classpath with wildcards * expanded. This is useful in situations where wildcard syntax isn't usable. * With the --jar argument, it writes the classpath as a manifest in a jar file. * This is useful in environments with short limitations on the maximum command * line length, where it may not be possible to specify the full classpath in a * command. For example, the maximum command line length on Windows is 8191 * characters. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | ComparableVersion.java | Code source of this file:| http://grepcode.com/file/repo1.maven.org/maven2/| org.apache.maven/maven-artifact/3.1.1/| org/apache/maven/artifact/versioning/ComparableVersion.java/|| Modifications made on top of the source:| 1. Changed| package org.apache.maven.artifact.versioning;| to| package org.apache.hadoop.util;| 2. Removed author tags to clear hadoop author tag warning|
* Generic implementation of version comparison.
* * <p>Features:
* <ul>
* <li>mixing of '<code>-</code>' (dash) and '<code>.</code>' (dot) separators,</li>
* <li>transition between characters and digits also constitutes a separator:
* <code>1.0alpha1 => [1, 0, alpha, 1]</code></li>
* <li>unlimited number of version components,</li>
* <li>version components in the text can be digits or strings,</li>
* <li>strings are checked for well-known qualifiers and the qualifier ordering is used for version ordering.
* Well-known qualifiers (case insensitive) are:<ul>
* <li><code>alpha</code> or <code>a</code></li>
* <li><code>beta</code> or <code>b</code></li>
* <li><code>milestone</code> or <code>m</code></li>
* <li><code>rc</code> or <code>cr</code></li>
* <li><code>snapshot</code></li>
* <li><code>(the empty string)</code> or <code>ga</code> or <code>final</code></li>
* <li><code>sp</code></li>
* </ul>
* Unknown qualifiers are considered after known qualifiers, with lexical order (always case insensitive),
* </li>
* <li>a dash usually precedes a qualifier, and is always less important than something preceded with a dot.</li>
* </ul><p>
*
* @see <a href="https://cwiki.apache.org/confluence/display/MAVENOLD/Versioning">"Versioning" on Maven Wiki</a>
|
* Represents a numeric item in the version item list.
|
* Represents a string in the version item list, usually a qualifier.
|
* Represents a version list item. This class is used both for the global item list and for sub-lists (which start
* with '-(number)' in the version specification). | Generic implementation of version comparison. Represents a numeric item in the version item list.
|
* Represents a string in the version item list, usually a qualifier.
|
* Represents a version list item. This class is used both for the global item list and for sub-lists (which start
* with '-(number)' in the version specification). | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | ConfigurationException.java | * Exception to throw in case of a configuration problem. | Exception to throw in case of a configuration problem. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | ContainerFinishData.java | * The class contains the fields that can be determined when
* <code>RMContainer</code> finishes, and that need to be stored persistently. | The class contains the fields that can be determined when * <code>RMContainer</code> finishes, and that need to be stored persistently. | null | The class contains the fields that can be determined when * <code>RMContainer</code> finishes, and that need to be stored persistently. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | CpuTimeTracker.java | * Utility for sampling and computing CPU usage. | Utility for sampling and computing CPU usage | null | Utility for sampling and computing CPU usage | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | DBProfile.java | * User visible configs based RocksDB tuning page. Documentation for Options.
* <p>
* https://github.com/facebook/rocksdb/blob/master/include/rocksdb/options.h
* <p>
* Most tuning parameters are based on this URL.
* <p>
* https://github.com/facebook/rocksdb/wiki/Setup-Options-and-Basic-Tuning | * User visible configs based RocksDB tuning page. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | DefaultAnonymizingRumenSerializer.java | * Default Rumen JSON serializer. | Default Rumen JSON serializer. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | DelegatingSSLSocketFactory.java | * A {@link SSLSocketFactory} that can delegate to various SSL implementations.
* Specifically, either OpenSSL or JSSE can be used. OpenSSL offers better
* performance than JSSE and is made available via the
* <a href="https://github.com/wildfly/wildfly-openssl">wildlfy-openssl</a>
* library.
*
* <p>
* The factory has several different modes of operation:
* <ul>
* <li>OpenSSL: Uses the wildly-openssl library to delegate to the
* system installed OpenSSL. If the wildfly-openssl integration is not
* properly setup, an exception is thrown.</li>
* <li>Default: Attempts to use the OpenSSL mode, if it cannot load the
* necessary libraries, it falls back to the Default_JSEE mode.</li>
* <li>Default_JSSE: Delegates to the JSSE implementation of SSL, but
* it disables the GCM cipher when running on Java 8.</li>
* <li>Default_JSSE_with_GCM: Delegates to the JSSE implementation of
* SSL with no modification to the list of enabled ciphers.</li>
* </ul>
* </p>
| This should only be modified within the #initializeDefaultFactory|
* Default indicates Ordered, preferred OpenSSL, if failed to load then fall
* back to Default_JSSE.
*
* <p>
* Default_JSSE is not truly the the default JSSE implementation because
* the GCM cipher is disabled when running on Java 8. However, the name
* was not changed in order to preserve backwards compatibility. Instead,
* a new mode called Default_JSSE_with_GCM delegates to the default JSSE
* implementation with no changes to the list of enabled ciphers.
* </p> | * A {@link SSLSocketFactory} that can delegate to various SSL implementations. | Specifically, either OpenSSL or JSSE can be used. OpenSSL offers better
* performance than JSSE and is made available via the
* <a href="https://github.com/wildfly/wildfly-openssl">wildlfy-openssl</a>
* library. | <p>
* Default_JSSE is not truly the the default JSSE implementation because
* the GCM cipher is disabled when running on Java 8. However, the name
* was not changed in order to preserve backwards compatibility. Instead,
* a new mode called Default_JSSE_with_GCM delegates to the default JSSE
* implementation with no changes to the list of enabled ciphers.
* </p> | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | DelegationTokenIdentifier.java | * A delegation token identifier that is specific to HDFS. | A delegation token identifier that is specific to HDFS. | A delegation token identifier that is specific to HDFS. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | DeleteApplicationHomeSubClusterRequest.java | * The request to <code>Federation state store</code> to delete the mapping of
* home subcluster of a submitted application. | The request to <code>Federation state store</code> to delete the mapping of * home subcluster of a submitted application. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | DFSConfigKeys.java | * This class contains constants for configuration keys and default values
* used in hdfs. | This class contains constants for configuration keys and default values used in hdfs. | null | This class contains constants for configuration keys and default values used in hdfs. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | DfsServlet.java | * A base class for the servlets in DFS. | A base class for the servlets in DFS | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | DiskBalancerCluster.java | * DiskBalancerCluster represents the nodes that we are working against.
* <p>
* Please Note :
* Semantics of inclusionList and exclusionLists.
* <p>
* If a non-empty inclusionList is specified then the diskBalancer assumes that
* the user is only interested in processing that list of nodes. This node list
* is checked against the exclusionList and only the nodes in inclusionList but
* not in exclusionList is processed.
* <p>
* if inclusionList is empty, then we assume that all live nodes in the nodes is
* to be processed by diskBalancer. In that case diskBalancer will avoid any
* nodes specified in the exclusionList but will process all nodes in the
* cluster.
* <p>
* In other words, an empty inclusionList is means all the nodes otherwise
* only a given list is processed and ExclusionList is always honored. | DiskBalancerCluster represents the nodes that we are working against. | Please Note :
* Semantics of inclusionList and exclusionLists.
* <p>
* If a non-empty inclusionList is specified then the diskBalancer assumes that
* the user is only interested in processing that list of nodes. This node list
* is checked against the exclusionList and only the nodes in inclusionList but
* not in exclusionList is processed.
* <p>
* if inclusionList is empty, then we assume that all live nodes in the nodes is
* to be processed by diskBalancer. In that case diskBalancer will avoid any
* nodes specified in the exclusionList but will process all nodes in the
* cluster.
* <p>
* In other words, an empty inclusionList is means all the nodes otherwise
* only a given list is processed and ExclusionList is always honored. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | DistributedSchedulingAllocateRequest.java | * Object used by the Application Master when distributed scheduling is enabled,
* in order to forward the {@link AllocateRequest} for GUARANTEED containers to
* the Resource Manager, and to notify the Resource Manager about the allocation
* of OPPORTUNISTIC containers through the Distributed Scheduler. | * Object used by the Application Master when distributed scheduling is enabled, | in order to forward the {@link AllocateRequest} for GUARANTEED containers to * the Resource Manager, and to notify the Resource Manager about the allocation * of OPPORTUNISTIC containers through the Distributed Scheduler. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | DockerKillCommand.java | * Encapsulates the docker kill command and its command line arguments. | Encapsulates the docker kill command and its command line arguments. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | EditLogTailer.java | * EditLogTailer represents a thread which periodically reads from edits
* journals and applies the transactions contained within to a given
* FSNamesystem.
|
* The thread which does the actual work of tailing edits journals and
* applying the transactions to the FSNS.
|
* Manage the 'active namenode proxy'. This cannot just be the a single proxy since we could
* failover across a number of NameNodes, rather than just between an active and a standby.
* <p>
* We - lazily - get a proxy to one of the configured namenodes and attempt to make the request
* against it. If it doesn't succeed, either because the proxy failed to be created or the request
* failed, we try the next NN in the list. We try this up to the configuration maximum number of
* retries before throwing up our hands. A working proxy is retained across attempts since we
* expect the active NameNode to switch rarely.
* <p>
* This mechanism is <b>very bad</b> for cases where we care about being <i>fast</i>; it just
* blindly goes and tries namenodes. | EditLogTailer represents a thread which periodically reads from edits
* journals and applies the transactions contained within to a given
* FSNamesystem. | |
* The thread which does the actual work of tailing edits journals and
* applying the transactions to the FSNS.
| | * Manage the 'active namenode proxy'. This cannot just be the a single proxy since we could
* failover across a number of NameNodes, rather than just between an active and a standby.
* <p>
* We - lazily - get a proxy to one of the configured namenodes and attempt to make the request
* against it. If it doesn't succeed, either because the proxy failed to be created or the request
* failed, we try the next NN in the list. We try this up to the configuration maximum number of
* retries before throwing up our hands. A working proxy is retained across attempts since we
* expect the active NameNode to switch rarely.
* <p>
* This mechanism is <b>very bad</b> for cases where we care about being <i>fast</i>; it just
* blindly goes and tries namenodes. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | ErasureCodingPolicyManager.java | * This manages erasure coding policies predefined and activated in the system.
* It loads customized policies and syncs with persisted ones in
* NameNode image.
*
* This class is instantiated by the FSNamesystem. | This manages erasure coding policies predefined and activated in the system | It loads customized policies and syncs with persisted ones in
* NameNode image. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | EventWatcher.java | * Event watcher the (re)send a message after timeout.
* <p>
* Event watcher will send the tracked payload/event after a timeout period
* unless a confirmation from the original event (completion event) is arrived.
*
* @param <TIMEOUT_PAYLOAD> The type of the events which are tracked.
* @param <COMPLETION_PAYLOAD> The type of event which could cancel the
* tracking. | * Event watcher the (re)send a message after timeout. | null | Event watcher will send the tracked payload/event after a timeout period
* unless a confirmation from the original event (completion event) is arrived. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | Expression.java | * Interface describing an expression to be used in the
* {@link org.apache.hadoop.fs.shell.find.Find} command. | Interface describing an expression to be used in the * {@link org.apache.hadoop.fs.shell.find.Find} command. | null | to be used in the * {@link org.apache.hadoop.fs.shell.find.Find} command. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | FairOrderingPolicy.java | * An OrderingPolicy which orders SchedulableEntities for fairness (see
* FairScheduler
* FairSharePolicy), generally, processes with lesser usage are lesser. If
* sizedBasedWeight is set to true then an application with high demand
* may be prioritized ahead of an application with less usage. This
* is to offset the tendency to favor small apps, which could result in
* starvation for large apps if many small ones enter and leave the queue
* continuously (optional, default false) | * An OrderingPolicy which orders SchedulableEntities for fairness (see
* FairScheduler
* FairSharePolicy), generally, processes with lesser usage are lesser | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | FederationPolicyException.java | * Generic policy exception. | Generic policy exception | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | FederationProtocolPBTranslator.java | * Helper class for setting/getting data elements in an object backed by a
* protobuf implementation. | Helper class for setting/getting data elements in an object backed by a * protobuf implementation. | null | for setting/getting data elements in an object backed by a * protobuf implementation. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | FederationStateStoreInvalidInputException.java | * Exception thrown by the {@code FederationMembershipStateStoreInputValidator},
* {@code FederationApplicationHomeSubClusterStoreInputValidator},
* {@code FederationPolicyStoreInputValidator} if the input is invalid.
* | Exception thrown by the {@code FederationMembershipStateStoreInputValidator}, | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | FileSystemApplicationHistoryStore.java | * File system implementation of {@link ApplicationHistoryStore}. In this
* implementation, one application will have just one file in the file system,
* which contains all the history data of one application, and its attempts and
* containers. {@link #applicationStarted(ApplicationStartData)} is supposed to
* be invoked first when writing any history data of one application and it will
* open a file, while {@link #applicationFinished(ApplicationFinishData)} is
* supposed to be last writing operation and will close the file. | * File system implementation of {@link ApplicationHistoryStore}. | In this * implementation, one application will have just one file in the file system, * which contains all the history data of one application, and its attempts and * containers. {@link #applicationStarted(ApplicationStartData)} is supposed to
* be invoked first when writing any history data of one application and it will
* open a file, while {@link #applicationFinished(ApplicationFinishData)} is
* supposed to be last writing operation and will close the file. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | FsConstants.java | * FileSystem related constants. | FileSystem related constants | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | GetApplicationHomeSubClusterResponsePBImpl.java | * Protocol buffer based implementation of
* {@link GetApplicationHomeSubClusterResponse}. | Protocol buffer based implementation of * {@link GetApplicationHomeSubClusterResponse}. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | GetNamespaceInfoResponse.java | * API response for listing HDFS namespaces present in the state store. | API response for listing HDFS namespaces present in the state store. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | GetNodesToAttributesResponse.java | * <p>
* The response sent by the <code>ResourceManager</code> to a client requesting
* nodes to attributes mapping.
* </p>
*
* @see ApplicationClientProtocol#getNodesToAttributes
* (GetNodesToAttributesRequest) | The response sent by the <code>ResourceManager</code> to a client requesting * nodes to attributes mapping. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | GetSafeModeRequestPBImpl.java | * Protobuf implementation of the state store API object
* GetSafeModeRequest. | Protobuf implementation of the state store API object * GetSafeModeRequest. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | GetSubClusterPolicyConfigurationRequestPBImpl.java | * Protocol buffer based implementation of
* {@link GetSubClusterPolicyConfigurationRequest}. | Protocol buffer based implementation of * {@link GetSubClusterPolicyConfigurationRequest}. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | HadoopIllegalArgumentException.java | * Indicates that a method has been passed illegal or invalid argument. This
* exception is thrown instead of IllegalArgumentException to differentiate the
* exception thrown in Hadoop implementation from the one thrown in JDK. | Indicates that a method has been passed illegal or invalid argument. | This * exception is thrown instead of IllegalArgumentException to differentiate the * exception thrown in Hadoop implementation from the one thrown in JDK. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | HashResolver.java | * Order the destinations based on consistent hashing. | Order the destinations based on consistent hashing. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | HttpHeaderConfigurations.java | * Responsible to keep all abfs http headers here. | null | Responsible to keep all abfs http headers here. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | IDataLoader.java | * an IDataLoader loads data on demand | an IDataLoader loads data on demand | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | InconsistentS3ClientFactory.java | * S3 Client factory used for testing with eventual consistency fault injection.
* This client is for testing <i>only</i>; it is in the production
* {@code hadoop-aws} module to enable integration tests to use this
* just by editing the Hadoop configuration used to bring up the client. | S3 Client factory used for testing with eventual consistency fault injection. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | InfoKeyHandler.java | * Executes Info Object. | Executes Info Object | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | InvalidContainerRequestException.java | * Thrown when an arguments are combined to construct a
* <code>AMRMClient.ContainerRequest</code> in an invalid way. | * Thrown when an arguments are combined to construct a * <code>AMRMClient.ContainerRequest</code> in an invalid way. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | ITestAbfsReadWriteAndSeek.java | * Test read, write and seek.
* Uses package-private methods in AbfsConfiguration, which is why it is in
* this package. | Test read, write and seek. | null | Uses package-private methods in AbfsConfiguration, which is why it is in * this package. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | ITestAzureNativeContractSeek.java | * Contract test. | Contract test | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | ITestCommitOperations.java | * Test the low-level binding of the S3A FS to the magic commit mechanism,
* and handling of the commit operations.
* This is done with an inconsistent client. | * Test the low-level binding of the S3A FS to the magic commit mechanism, * and handling of the commit operations. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | ITestListPerformance.java | * Test list performance. | Test list performance | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | ITestS3Select.java | * Test the S3 Select feature with some basic SQL Commands.
* Executed if the destination store declares its support for the feature. | Test the S3 Select feature with some basic SQL Commands. | Executed if the destination store declares its support for the feature. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | ITestS3SelectCLI.java | * Test the S3 Select CLI through some operations against landsat
* and files generated from it. | Test the S3 Select CLI through some operations against landsat * and files generated from it. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | KerberosDelegationTokenAuthenticator.java | * The <code>KerberosDelegationTokenAuthenticator</code> provides support for
* Kerberos SPNEGO authentication mechanism and support for Hadoop Delegation
* Token operations.
* <p>
* It falls back to the {@link PseudoDelegationTokenAuthenticator} if the HTTP
* endpoint does not trigger a SPNEGO authentication | The <code>KerberosDelegationTokenAuthenticator</code> provides support for
* Kerberos SPNEGO authentication mechanism and support for Hadoop Delegation
* Token operations. | It falls back to the {@link PseudoDelegationTokenAuthenticator} if the HTTP
* endpoint does not trigger a SPNEGO authentication | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | LocalizationStatusPBImpl.java | * PB Impl of {@link LocalizationStatus}. | PB Impl of {@link LocalizationStatus}. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | LocatedFileStatus.java | * This class defines a FileStatus that includes a file's block locations. | This class defines a FileStatus that includes a file's block locations | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | LoggedTask.java | * A {@link LoggedTask} represents a [hadoop] task that is part of a hadoop job.
* It knows about the [pssibly empty] sequence of attempts, its I/O footprint,
* and its runtime.
* * All of the public methods are simply accessors for the instance variables we
* want to write out in the JSON files.
* | A {@link LoggedTask} represents a [hadoop] task that is part of a hadoop job. | It knows about the [pssibly empty] sequence of attempts, its I/O footprint,
* and its runtime. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | LogParserUtil.java | * Common utility functions for {@link LogParser}. | Common utility functions for {@link LogParser} | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | LogWebService.java | * Support only ATSv2 client only. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | LRUCacheHashMap.java | * LRU cache with a configurable maximum cache size and access order. | LRU cache with a configurable maximum cache size and access order. | LRU cache with a configurable maximum cache size and access order. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | MapContext.java | * The context that is given to the {@link Mapper}.
* @param <KEYIN> the key input type to the Mapper
* @param <VALUEIN> the value input type to the Mapper
* @param <KEYOUT> the key output type from the Mapper
* @param <VALUEOUT> the value output type from the Mapper | The context that is given to the {@link Mapper}. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | MetaBlockAlreadyExists.java | * Exception - Meta Block with the same name already exists. | Exception - Meta Block with the same name already exists. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | MetricsCache.java | * A metrics cache for sinks that don't support sparse updates.
|
* Cached record | A metrics cache for sinks that don't support sparse updates. Context of the Queues in Scheduler. | Cached record | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | NativeBatchProcessor.java | * used to create channel, transfer data and command between Java and native | null | null | used to create channel, transfer data and command between Java and native | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | NativeSingleLineParser.java | * This sample parser will parse the sample log and extract the resource
* skyline.
* <p> The expected log format is: NormalizedJobName NumInstances SubmitTime
* StartTime EndTime JobInstanceName memUsage coreUsage | This sample parser will parse the sample log and extract the resource
* skyline. | <p> The expected log format is: NormalizedJobName NumInstances SubmitTime
* StartTime EndTime JobInstanceName memUsage coreUsage | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | Nfs3Metrics.java | * This class is for maintaining the various NFS gateway activity statistics and
* publishing them through the metrics interfaces. | This class is for maintaining the various NFS gateway activity statistics and * publishing them through the metrics interfaces. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | Nfs3Status.java | * Success or error status is reported in NFS3 responses. | * Success or error status is reported in NFS3 responses. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | Node2ObjectsMap.java | * This data structure maintains the list of containers that is on a datanode.
* This information is built from the DN container reports. | This data structure maintains the list of containers that is on a datanode. | This information is built from the DN container reports. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | NodeUpdateType.java | * <p>Taxonomy of the <code>NodeState</code> that a
* <code>Node</code> might transition into.</p>
* | <p>Taxonomy of the <code>NodeState</code> that a * <code>Node</code> might transition into.</p> | <p>Taxonomy of the <code>NodeState</code> that a * <code>Node</code> might transition into.</p> | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | NullOutputFormat.java | * Consume all outputs and put them in /dev/null. | Consume all outputs and put them in /dev/null. | Consume all outputs and put them in /dev/null. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | OMNodeDetails.java | * This class stores OM node details.
|
* Builder class for OMNodeDetails. | * This class stores OM node details.
|
* Builder class for OMNodeDetails. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | OpportunisticContainersStatusPBImpl.java | * Protocol Buffer implementation of OpportunisticContainersStatus. | Protocol Buffer implementation of OpportunisticContainersStatus | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | OzoneObj.java | * Class representing an unique ozone object.
* |
* Ozone Objects supported for ACL.
|
* Ozone Objects supported for ACL. | Class representing an unique ozone object | Ozone Objects supported for ACL. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | Parser.java | A class for parsing outputs | A class for parsing outputs | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | PartialOutputCommitter.java | * Interface for an {@link org.apache.hadoop.mapreduce.OutputCommitter}
* implementing partial commit of task output, as during preemption. | * Interface for an {@link org.apache.hadoop.mapreduce.OutputCommitter} * implementing partial commit of task output, as during preemption. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | PartitionResourcesInfo.java | * This class represents queue/user resource usage info for a given partition | This class represents queue/user resource usage info for a given partition | null | This class represents queue/user resource usage info for a given partition | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | PlanningQuotaException.java | * This exception is thrown if the user quota is exceed while accepting or
* updating a reservation. | This exception is thrown if the user quota is exceed while accepting or * updating a reservation. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | ProcessIdFileReader.java | * Helper functionality to read the pid from a file. | Helper functionality to read the pid from a file. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | QuasiMonteCarlo.java | * A map/reduce program that estimates the value of Pi
* using a quasi-Monte Carlo (qMC) method.
* Arbitrary integrals can be approximated numerically by qMC methods.
* In this example,
* we use a qMC method to approximate the integral $I = \int_S f(x) dx$,
* where $S=[0,1)^2$ is a unit square,
* $x=(x_1,x_2)$ is a 2-dimensional point,
* and $f$ is a function describing the inscribed circle of the square $S$,
* $f(x)=1$ if $(2x_1-1)^2+(2x_2-1)^2 <= 1$ and $f(x)=0$, otherwise.
* It is easy to see that Pi is equal to $4I$.
* So an approximation of Pi is obtained once $I$ is evaluated numerically.
* * There are better methods for computing Pi.
* We emphasize numerical approximation of arbitrary integrals in this example.
* For computing many digits of Pi, consider using bbp.
*
* The implementation is discussed below.
*
* Mapper:
* Generate points in a unit square
* and then count points inside/outside of the inscribed circle of the square.
*
* Reducer:
* Accumulate points inside/outside results from the mappers.
*
* Let numTotal = numInside + numOutside.
* The fraction numInside/numTotal is a rational approximation of
* the value (Area of the circle)/(Area of the square) = $I$,
* where the area of the inscribed circle is Pi/4
* and the area of unit square is 1.
* Finally, the estimated value of Pi is 4(numInside/numTotal). | 2-dimensional Halton sequence {H(i)},
* where H(i) is a 2-dimensional point and i >= 1 is the index.
* Halton sequence is used to generate sample points for Pi estimation. |
* Mapper class for Pi estimation.
* Generate points in a unit square
* and then count points inside/outside of the inscribed circle of the square.
|
* Reducer class for Pi estimation.
* Accumulate points inside/outside results from the mappers. | * A map/reduce program that estimates the value of Pi. | using a quasi-Monte Carlo (qMC) method. Arbitrary integrals can be approximated numerically by qMC methods. * There are better methods for computing Pi.
* We emphasize numerical approximation of arbitrary integrals in this example.
* For computing many digits of Pi, consider using bbp.
*
* The implementation is discussed below.
*
* Mapper:
* Generate points in a unit square
* and then count points inside/outside of the inscribed circle of the square.
*
* Reducer:
* Accumulate points inside/outside results from the mappers. * Mapper class for Pi estimation.
* Generate points in a unit square
* and then count points inside/outside of the inscribed circle of the square.
|
* Reducer class for Pi estimation.
* Accumulate points inside/outside results from the mappers. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | Query.java | * Check if a record matches a query. The query is usually a partial record.
*
* @param <T> Type of the record to query. | Check if a record matches a query. | The query is usually a partial record. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | QueueName.java | * Represents a queue name. | Represents a queue name. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | RandomKeyGenerator.java | * Data generator tool to generate as much keys as possible.
|
* Wrapper to hold ozone keyValidate entry.
|
* Validates the write done in ozone cluster. | Data generator tool to generate as much keys as possible. | Wrapper to hold ozone keyValidate entry.
|
* Validates the write done in ozone cluster. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | ReencryptionUpdater.java | * Class for finalizing re-encrypt EDEK operations, by updating file xattrs with
* edeks returned from reencryption.
* <p>
* The tasks are submitted by ReencryptionHandler.
* <p>
* It is assumed only 1 Updater will be running, since updating file xattrs
* requires namespace write lock, and performance gain from multi-threading
* is limited.
|
* Class to track re-encryption submissions of a single zone. It contains
* all the submitted futures, and statistics about how far the futures are
* processed.
|
* Class representing the task for one batch of a re-encryption command. It
* also contains statistics about how far this single batch has been executed.
|
* Class that encapsulates re-encryption details of a file. | Class for finalizing re-encrypt EDEK operations. Class representing the task for one batch of a re-encryption command. It
* also contains statistics about how far this single batch has been executed. Class that encapsulates re-encryption details of a file. | by updating file xattrs with * edeks returned from reencryption. * <p> * The tasks are submitted by ReencryptionHandler. * <p>. It is assumed only 1 Updater will be running, since updating file xattrs
* requires namespace write lock, and performance gain from multi-threading
* is limited.
|
* Class to track re-encryption submissions of a single zone. It contains
* all the submitted futures, and statistics about how far the futures are
* processed. It contains the
* file inode, stores the initial edek of the file, and the new edek
* after re-encryption.
* <p>
* Assumptions are the object initialization happens when dir lock is held,
* and inode is valid and is encrypted during initialization.
* <p>
* Namespace changes may happen during re-encryption, and if inode is changed
* the re-encryption is skipped. | Class for finalizing re-encrypt EDEK operations | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | RegistryInternalConstants.java | * Internal constants for the registry.
*
* These are the things which aren't visible to users.
* | Internal constants for the registry. | null | These are the things which aren't visible to users. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | RegistryOperations.java | * Registry Operations | Registry Operations | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | ReInitializeContainerRequestPBImpl.java | CHECKSTYLE:OFF | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | ResourceBlacklistRequest.java | * {@link ResourceBlacklistRequest} encapsulates the list of resource-names * which should be added or removed from the <em>blacklist</em> of resources * for the application.
* * @see ResourceRequest
* @see ApplicationMasterProtocol#allocate(org.apache.hadoop.yarn.api.protocolrecords.AllocateRequest) | * {@link ResourceBlacklistRequest} encapsulates the list of resource-names * which should be added or removed from the <em>blacklist</em> of resources * for the application. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | ResourceRequestsJsonVerifications.java | * Performs value verifications on
* {@link org.apache.hadoop.yarn.server.resourcemanager.webapp.dao.ResourceRequestInfo}
* objects against the values of {@link ResourceRequest}. With the help of the
* {@link Builder}, users can also make verifications of the custom resource
* types and its values.
|
* Builder class for {@link ResourceRequestsJsonVerifications}. | * Performs value verifications on
* {@link org.apache.hadoop.yarn.server.resourcemanager.webapp.dao.ResourceRequestInfo}
* objects against the values of {@link ResourceRequest}.
|
* Builder class for {@link ResourceRequestsJsonVerifications}. | With the help of the
* {@link Builder}, users can also make verifications of the custom resource
* types and its values. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | RetriableDirectoryCreateCommand.java | * This class extends Retriable command to implement the creation of directories
* with retries on failure. | This class extends Retriable command to implement the creation of directories * with retries on failure. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | RMAdminRequestInterceptor.java | * Defines the contract to be implemented by the request intercepter classes,
* that can be used to intercept and inspect messages sent from the client to
* the resource manager. | Defines the contract to be implemented by the request intercepter classes | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | RSLegacyRawErasureCoderFactory.java | * A raw coder factory for the legacy raw Reed-Solomon coder in Java. | A raw coder factory for the legacy raw Reed-Solomon coder in Java. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | SafeModeException.java | * This exception is thrown when the name node is in safe mode.
* Client cannot modified namespace until the safe mode is off.
* | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | SchedulerQueueManager.java | *
* Context of the Queues in Scheduler.
* | Context of the Queues in Scheduler. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
1 | SequenceFileRecordReader.java | * An {@link RecordReader} for {@link SequenceFile}s. | * An {@link RecordReader} for {@link SequenceFile}s. | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
Dataset Card for Code Comment Classification
Dataset Summary
The dataset contains class comments extracted from various big and diverse open-source projects of three programming languages Java, Smalltalk, and Python.
Supported Tasks and Leaderboards
Single-label text classification and Multi-label text classification
Languages
Java, Python, Smalltalk
Dataset Structure
Data Instances
{
"class" : "Absy.java",
"comment":"* Azure Blob File System implementation of AbstractFileSystem. * This impl delegates to the old FileSystem",
"summary":"Azure Blob File System implementation of AbstractFileSystem.",
"expand":"This impl delegates to the old FileSystem",
"rational":"",
"deprecation":"",
"usage":"",
"exception":"",
"todo":"",
"incomplete":"",
"commentedcode":"",
"directive":"",
"formatter":"",
"license":"",
"ownership":"",
"pointer":"",
"autogenerated":"",
"noise":"",
"warning":"",
"recommendation":"",
"precondition":"",
"codingGuidelines":"",
"extension":"",
"subclassexplnation":"",
"observation":"",
}
Data Fields
class: name of the class with the language extension.
comment: class comment of the class
categories: a category that sentence is classified to. It indicated a particular type of information.
Data Splits
10-fold cross validation
Dataset Creation
Curation Rationale
To identify the infomation embedded in the class comments across various projects and programming languages.
Source Data
Initial Data Collection and Normalization
It contains the dataset extracted from various open-source projects of three programming languages Java, Smalltalk, and Python.
Java
Each file contains all the extracted class comments from one project. We have a total of six java projects. We chose a sample of 350 comments from all these files for our experiment.
Eclipse.csv - Extracted class comments from the Eclipse project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub Eclipse.
Guava.csv - Extracted class comments from the Guava project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub Guava.
Guice.csv - Extracted class comments from the Guice project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub Guice.
Hadoop.csv - Extracted class comments from the Hadoop project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub Apache Hadoop
Spark.csv - Extracted class comments from the Apache Spark project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub Apache Spark
Vaadin.csv - Extracted class comments from the Vaadin project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub Vaadin
Parser_Details.md - Details of the parser used to parse class comments of Java Projects
Smalltalk/
Each file contains all the extracted class comments from one project. We have a total of seven Pharo projects. We chose a sample of 350 comments from all these files for our experiment.
GToolkit.csv - Extracted class comments from the GToolkit project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo.
Moose.csv - Extracted class comments from the Moose project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo.
PetitParser.csv - Extracted class comments from the PetitParser project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo.
Pillar.csv - Extracted class comments from the Pillar project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo.
PolyMath.csv - Extracted class comments from the PolyMath project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo.
Roassal2.csv -Extracted class comments from the Roassal2 project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo.
Seaside.csv - Extracted class comments from the Seaside project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo.
Parser_Details.md - Details of the parser used to parse class comments of Pharo Projects
Python/
Each file contains all the extracted class comments from one project. We have a total of seven Python projects. We chose a sample of 350 comments from all these files for our experiment.
Django.csv - Extracted class comments from the Django project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub Django
IPython.csv - Extracted class comments from the Ipython project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHubIPython
Mailpile.csv - Extracted class comments from the Mailpile project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub Mailpile
Pandas.csv - Extracted class comments from the Pandas project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub pandas
Pipenv.csv - Extracted class comments from the Pipenv project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub Pipenv
Pytorch.csv - Extracted class comments from the Pytorch project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub PyTorch
Requests.csv - Extracted class comments from the Requests project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub Requests
Parser_Details.md - Details of the parser used to parse class comments of Python Projects
Annotations
Annotation process
Four evaluators (all authors of this paper (https://doi.org/10.1016/j.jss.2021.111047)), each having at least four years of programming experience, participated in the annonation process. We partitioned Java, Python, and Smalltalk comments equally among all evaluators based on the distribution of the language's dataset to ensure the inclusion of comments from all projects and diversified lengths. Each classification is reviewed by three evaluators. The details are given in the paper Rani et al., JSS, 2021
Who are the annotators?
Personal and Sensitive Information
Author information embedded in the text
Additional Information
Dataset Curators
[Pooja Rani, Ivan, Manuel]
Licensing Information
[license: cc-by-nc-sa-4.0]
Citation Information
@article{RANI2021111047,
title = {How to identify class comment types? A multi-language approach for class comment classification},
journal = {Journal of Systems and Software},
volume = {181},
pages = {111047},
year = {2021},
issn = {0164-1212},
doi = {https://doi.org/10.1016/j.jss.2021.111047},
url = {https://www.sciencedirect.com/science/article/pii/S0164121221001448},
author = {Pooja Rani and Sebastiano Panichella and Manuel Leuenberger and Andrea {Di Sorbo} and Oscar Nierstrasz},
keywords = {Natural language processing technique, Code comment analysis, Software documentation}
}
- Downloads last month
- 53