Dataset Preview
Full Screen
The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
The dataset generation failed because of a cast error
Error code:   DatasetGenerationCastError
Exception:    DatasetGenerationCastError
Message:      An error occurred while generating the dataset

All the data files must have the same columns, but at some point there are 10 new columns ({'Parameters', 'DevelopmentNotes', 'Version', 'Todo', 'Exception', 'Warnings', 'Links', 'Dependecies', 'Summary', 'Usage'}) and 15 missing columns ({'usage', 'Incomplete', 'exception', 'summary', 'Commentedcode', 'todo', 'Ownership', 'Pointer', 'rational', 'License', 'deprecation', 'Autogenerated', 'formatter', 'directive', 'Warning'}).

This happened while the csv dataset builder was generating data using

hf://datasets/poojaruhal/Code-comment-classification/python_0_raw.csv (at revision 3d2bbff4d30d5c41d2cbf5b1d55fbc8d10cfdbaa)

Please either edit the data files to have matching columns, or separate them into different configurations (see docs at https://hf.co/docs/hub/datasets-manual-configuration#multiple-configurations)
Traceback:    Traceback (most recent call last):
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 2011, in _prepare_split_single
                  writer.write_table(table)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 585, in write_table
                  pa_table = table_cast(pa_table, self._schema)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2302, in table_cast
                  return cast_table_to_schema(table, schema)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2256, in cast_table_to_schema
                  raise CastError(
              datasets.table.CastError: Couldn't cast
              stratum: int64
              class: string
              comment: string
              Summary: string
              Usage: string
              Parameters: string
              Expand: string
              Version: string
              DevelopmentNotes: string
              Todo: string
              Exception: string
              Links: string
              Noise: string
              Warnings: string
              Recommendation: string
              Dependecies: string
              Precondition: string
              CodingGuidelines: string
              Extension: string
              Subclassexplnation: string
              Observation: string
              -- schema metadata --
              pandas: '{"index_columns": [{"kind": "range", "name": null, "start": 0, "' + 2742
              to
              {'stratum': Value(dtype='int64', id=None), 'class': Value(dtype='string', id=None), 'comment': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None), 'Expand': Value(dtype='string', id=None), 'rational': Value(dtype='string', id=None), 'deprecation': Value(dtype='string', id=None), 'usage': Value(dtype='string', id=None), 'exception': Value(dtype='string', id=None), 'todo': Value(dtype='string', id=None), 'Incomplete': Value(dtype='string', id=None), 'Commentedcode': Value(dtype='string', id=None), 'directive': Value(dtype='string', id=None), 'formatter': Value(dtype='string', id=None), 'License': Value(dtype='float64', id=None), 'Ownership': Value(dtype='string', id=None), 'Pointer': Value(dtype='string', id=None), 'Autogenerated': Value(dtype='string', id=None), 'Noise': Value(dtype='string', id=None), 'Warning': Value(dtype='string', id=None), 'Recommendation': Value(dtype='string', id=None), 'Precondition': Value(dtype='string', id=None), 'CodingGuidelines': Value(dtype='float64', id=None), 'Extension': Value(dtype='float64', id=None), 'Subclassexplnation': Value(dtype='string', id=None), 'Observation': Value(dtype='string', id=None)}
              because column names don't match
              
              During handling of the above exception, another exception occurred:
              
              Traceback (most recent call last):
                File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1321, in compute_config_parquet_and_info_response
                  parquet_operations = convert_to_parquet(builder)
                File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 935, in convert_to_parquet
                  builder.download_and_prepare(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1027, in download_and_prepare
                  self._download_and_prepare(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1122, in _download_and_prepare
                  self._prepare_split(split_generator, **prepare_split_kwargs)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1882, in _prepare_split
                  for job_id, done, content in self._prepare_split_single(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 2013, in _prepare_split_single
                  raise DatasetGenerationCastError.from_cast_error(
              datasets.exceptions.DatasetGenerationCastError: An error occurred while generating the dataset
              
              All the data files must have the same columns, but at some point there are 10 new columns ({'Parameters', 'DevelopmentNotes', 'Version', 'Todo', 'Exception', 'Warnings', 'Links', 'Dependecies', 'Summary', 'Usage'}) and 15 missing columns ({'usage', 'Incomplete', 'exception', 'summary', 'Commentedcode', 'todo', 'Ownership', 'Pointer', 'rational', 'License', 'deprecation', 'Autogenerated', 'formatter', 'directive', 'Warning'}).
              
              This happened while the csv dataset builder was generating data using
              
              hf://datasets/poojaruhal/Code-comment-classification/python_0_raw.csv (at revision 3d2bbff4d30d5c41d2cbf5b1d55fbc8d10cfdbaa)
              
              Please either edit the data files to have matching columns, or separate them into different configurations (see docs at https://hf.co/docs/hub/datasets-manual-configuration#multiple-configurations)

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

stratum
int64
class
string
comment
string
summary
string
Expand
string
rational
string
deprecation
null
usage
null
exception
null
todo
null
Incomplete
null
Commentedcode
null
directive
null
formatter
null
License
null
Ownership
null
Pointer
null
Autogenerated
null
Noise
null
Warning
null
Recommendation
null
Precondition
null
CodingGuidelines
null
Extension
null
Subclassexplnation
null
Observation
null
1
Abfss.java
* Azure Blob File System implementation of AbstractFileSystem. * This impl delegates to the old FileSystem
Azure Blob File System implementation of AbstractFileSystem.
This impl delegates to the old FileSystem
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
AbstractContractGetFileStatusTest.java
* Test getFileStatus and related listing operations. | the tree parameters. Kept small to avoid killing object store test| runs too much.| * Accept everything. | the tree parameters. Kept small to avoid killing object store test| runs too much.| * Accept nothing. | the tree parameters. Kept small to avoid killing object store test| runs too much.| * Path filter which only expects paths whose final name element * equals the {@code match} field. | the tree parameters. Kept small to avoid killing object store test| runs too much.| * A filesystem filter which exposes the protected method * {@link #listLocatedStatus(Path, PathFilter)}.
Test getFileStatus and related listing operations.
| the tree parameters. Kept small to avoid killing object store test| runs too much.| * Accept everything. | the tree parameters. Kept small to avoid killing object store test| runs too much.| * Accept nothing. | the tree parameters. Kept small to avoid killing object store test| runs too much.| * Path filter which only expects paths whose final name element * equals the {@code match} field. | the tree parameters. Kept small to avoid killing object store test| runs too much.| * A filesystem filter which exposes the protected method * {@link #listLocatedStatus(Path, PathFilter)}.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
AbstractFSContract.java
* Class representing a filesystem contract that a filesystem * implementation is expected implement. * * Part of this contract class is to allow FS implementations to * provide specific opt outs and limits, so that tests can be * skip unsupported features (e.g. case sensitivity tests), * dangerous operations (e.g. trying to delete the root directory), * and limit filesize and other numeric variables for scale tests
Class representing a filesystem contract that a filesystem * implementation is expected implement.
null
Part of this contract class is to allow FS implementations to * provide specific opt outs and limits, so that tests can be * skip unsupported features (e.g. case sensitivity tests), * dangerous operations (e.g. trying to delete the root directory), * and limit filesize and other numeric variables for scale tests
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
AbstractS3ACommitterFactory.java
* Dynamically create the output committer based on subclass type and settings.
null
Dynamically create the output committer based on subclass type and settings.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
AbstractTFLaunchCommandTestHelper.java
* This class is an abstract base class for testing Tensorboard and TensorFlow * launch commands.
This class is an abstract base class for testing Tensorboard and TensorFlow and launch commands.
null
This class is an abstract base class for testing Tensorboard and TensorFlow and launch commands.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
ApplicationConstants.java
* This is the API for the applications comprising of constants that YARN sets * up for the applications and the containers. * * TODO: Investigate the semantics and security of each cross-boundary refs. | * The type of launch for the container. | * Environment for Applications. * * Some of the environment variables for applications are <em>final</em> * i.e. they cannot be modified by the applications.
* This is the API for the applications comprising of constants that YARN sets * up for the applications and the containers.
The type of launch for the container. | * Environment for Applications. * * Some of the environment variables for applications are <em>final</em> * i.e. they cannot be modified by the applications.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
ApplicationStateData.java
* Contains all the state data that needs to be stored persistently * for an Application
Contains all the state data that needs to be stored persistently * for an Application
null
needs to be stored persistently * for an Application
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
AutoInputFormat.java
* An {@link InputFormat} that tries to deduce the types of the input files * automatically. It can currently handle text and sequence files.
An {@link InputFormat} that tries to deduce the types of the input files * automatically.
It can currently handle text and sequence files.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
BalancingPolicy.java
* Balancing policy. * Since a datanode may contain multiple block pools, * {@link Pool} implies {@link Node} * but NOT the other way around | * Cluster is balanced if each node is balanced. | * Cluster is balanced if each pool in each node is balanced.
* Balancing policy.
Since a datanode may contain multiple block pools, * {@link Pool} implies {@link Node} * but NOT the other way around |
Cluster is balanced if each node is balanced. | * Cluster is balanced if each pool in each node is balanced.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
BaseRouterWebServicesTest.java
* Base class for all the RouterRMAdminService test cases. It provides utility * methods that can be used by the concrete test case classes. *
* Base class for all the RouterRMAdminService test cases.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
BatchedRequests.java
* A grouping of Scheduling Requests which are sent to the PlacementAlgorithm * to place as a batch. The placement algorithm tends to give more optimal * placements if more requests are batched together. | PlacementAlgorithmOutput attempt - the number of times the requests in this| * Iterator Type.
A grouping of Scheduling Requests which are sent to the PlacementAlgorithm * to place as a batch Iterator Type.
The placement algorithm tends to give more optimal * placements if more requests are batched together.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
BlockPlacementStatusWithNodeGroup.java
* An implementation of @see BlockPlacementStatus for * @see BlockPlacementPolicyWithNodeGroup
An implementation of @see BlockPlacementStatus
null
for * @see BlockPlacementPolicyWithNodeGroup
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
BlocksMap.java
* This class maintains the map from a block to its metadata. * block's metadata currently includes blockCollection it belongs to and * the datanodes that store the block.
This class maintains the map from a block to its metadata.
block's metadata currently includes blockCollection it belongs to and * the datanodes that store the block.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
BlockUtils.java
* Utils functions to help block functions.
Utils functions to help block functions
null
Utils functions to help block functions
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
ByteArrayEncodingState.java
* A utility class that maintains encoding state during an encode call using * byte array inputs.
A utility class that maintains encoding state during an encode call using * byte array inputs.
A utility class that maintains encoding state during an encode call using * byte array inputs.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
CapacitySchedulerPlanFollower.java
* This class implements a {@link PlanFollower}. This is invoked on a timer, and * it is in charge to publish the state of the {@link Plan}s to the underlying * {@link CapacityScheduler}. This implementation does so, by * adding/removing/resizing leaf queues in the scheduler, thus affecting the * dynamic behavior of the scheduler in a way that is consistent with the * content of the plan. It also updates the plan's view on how much resources * are available in the cluster. * * This implementation of PlanFollower is relatively stateless, and it can * synchronize schedulers and Plans that have arbitrary changes (performing set * differences among existing queues). This makes it resilient to frequency of * synchronization, and RM restart issues (no "catch up" is necessary).
This class implements a {@link PlanFollower}. This is invoked on a timer, and * it is in charge to publish the state of the {@link Plan}s to the underlying * {@link CapacityScheduler}.
This implementation does so, by * adding/removing/resizing leaf queues in the scheduler, thus affecting the * dynamic behavior of the scheduler in a way that is consistent with the * content of the plan. It also updates the plan's view on how much resources * are available in the cluster.
This implementation of PlanFollower is relatively stateless, and it can * synchronize schedulers and Plans that have arbitrary changes (performing set * differences among existing queues). This makes it resilient to frequency of * synchronization, and RM restart issues (no "catch up" is necessary).
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
Classpath.java
* Command-line utility for getting the full classpath needed to launch a Hadoop * client application. If the hadoop script is called with "classpath" as the * command, then it simply prints the classpath and exits immediately without * launching a JVM. The output likely will include wildcards in the classpath. * If there are arguments passed to the classpath command, then this class gets * called. With the --glob argument, it prints the full classpath with wildcards * expanded. This is useful in situations where wildcard syntax isn't usable. * With the --jar argument, it writes the classpath as a manifest in a jar file. * This is useful in environments with short limitations on the maximum command * line length, where it may not be possible to specify the full classpath in a * command. For example, the maximum command line length on Windows is 8191 * characters.
Command-line utility for getting the full classpath needed to launch a Hadoop * client application.
If the hadoop script is called with "classpath" as the * command, then it simply prints the classpath and exits immediately without * launching a JVM. The output likely will include wildcards in the classpath. * If there are arguments passed to the classpath command, then this class gets * called. With the --glob argument, it prints the full classpath with wildcards * expanded. This is useful in situations where wildcard syntax isn't usable. * With the --jar argument, it writes the classpath as a manifest in a jar file. * This is useful in environments with short limitations on the maximum command * line length, where it may not be possible to specify the full classpath in a * command. For example, the maximum command line length on Windows is 8191 * characters.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
ComparableVersion.java
Code source of this file:| http://grepcode.com/file/repo1.maven.org/maven2/| org.apache.maven/maven-artifact/3.1.1/| org/apache/maven/artifact/versioning/ComparableVersion.java/|| Modifications made on top of the source:| 1. Changed| package org.apache.maven.artifact.versioning;| to| package org.apache.hadoop.util;| 2. Removed author tags to clear hadoop author tag warning| * Generic implementation of version comparison. * * <p>Features: * <ul> * <li>mixing of '<code>-</code>' (dash) and '<code>.</code>' (dot) separators,</li> * <li>transition between characters and digits also constitutes a separator: * <code>1.0alpha1 =&gt; [1, 0, alpha, 1]</code></li> * <li>unlimited number of version components,</li> * <li>version components in the text can be digits or strings,</li> * <li>strings are checked for well-known qualifiers and the qualifier ordering is used for version ordering. * Well-known qualifiers (case insensitive) are:<ul> * <li><code>alpha</code> or <code>a</code></li> * <li><code>beta</code> or <code>b</code></li> * <li><code>milestone</code> or <code>m</code></li> * <li><code>rc</code> or <code>cr</code></li> * <li><code>snapshot</code></li> * <li><code>(the empty string)</code> or <code>ga</code> or <code>final</code></li> * <li><code>sp</code></li> * </ul> * Unknown qualifiers are considered after known qualifiers, with lexical order (always case insensitive), * </li> * <li>a dash usually precedes a qualifier, and is always less important than something preceded with a dot.</li> * </ul><p> * * @see <a href="https://cwiki.apache.org/confluence/display/MAVENOLD/Versioning">"Versioning" on Maven Wiki</a> | * Represents a numeric item in the version item list. | * Represents a string in the version item list, usually a qualifier. | * Represents a version list item. This class is used both for the global item list and for sub-lists (which start * with '-(number)' in the version specification).
Generic implementation of version comparison. Represents a numeric item in the version item list. | * Represents a string in the version item list, usually a qualifier. | * Represents a version list item. This class is used both for the global item list and for sub-lists (which start * with '-(number)' in the version specification).
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
ConfigurationException.java
* Exception to throw in case of a configuration problem.
Exception to throw in case of a configuration problem.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
ContainerFinishData.java
* The class contains the fields that can be determined when * <code>RMContainer</code> finishes, and that need to be stored persistently.
The class contains the fields that can be determined when * <code>RMContainer</code> finishes, and that need to be stored persistently.
null
The class contains the fields that can be determined when * <code>RMContainer</code> finishes, and that need to be stored persistently.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
CpuTimeTracker.java
* Utility for sampling and computing CPU usage.
Utility for sampling and computing CPU usage
null
Utility for sampling and computing CPU usage
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
DBProfile.java
* User visible configs based RocksDB tuning page. Documentation for Options. * <p> * https://github.com/facebook/rocksdb/blob/master/include/rocksdb/options.h * <p> * Most tuning parameters are based on this URL. * <p> * https://github.com/facebook/rocksdb/wiki/Setup-Options-and-Basic-Tuning
* User visible configs based RocksDB tuning page.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
DefaultAnonymizingRumenSerializer.java
* Default Rumen JSON serializer.
Default Rumen JSON serializer.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
DelegatingSSLSocketFactory.java
* A {@link SSLSocketFactory} that can delegate to various SSL implementations. * Specifically, either OpenSSL or JSSE can be used. OpenSSL offers better * performance than JSSE and is made available via the * <a href="https://github.com/wildfly/wildfly-openssl">wildlfy-openssl</a> * library. * * <p> * The factory has several different modes of operation: * <ul> * <li>OpenSSL: Uses the wildly-openssl library to delegate to the * system installed OpenSSL. If the wildfly-openssl integration is not * properly setup, an exception is thrown.</li> * <li>Default: Attempts to use the OpenSSL mode, if it cannot load the * necessary libraries, it falls back to the Default_JSEE mode.</li> * <li>Default_JSSE: Delegates to the JSSE implementation of SSL, but * it disables the GCM cipher when running on Java 8.</li> * <li>Default_JSSE_with_GCM: Delegates to the JSSE implementation of * SSL with no modification to the list of enabled ciphers.</li> * </ul> * </p> | This should only be modified within the #initializeDefaultFactory| * Default indicates Ordered, preferred OpenSSL, if failed to load then fall * back to Default_JSSE. * * <p> * Default_JSSE is not truly the the default JSSE implementation because * the GCM cipher is disabled when running on Java 8. However, the name * was not changed in order to preserve backwards compatibility. Instead, * a new mode called Default_JSSE_with_GCM delegates to the default JSSE * implementation with no changes to the list of enabled ciphers. * </p>
* A {@link SSLSocketFactory} that can delegate to various SSL implementations.
Specifically, either OpenSSL or JSSE can be used. OpenSSL offers better * performance than JSSE and is made available via the * <a href="https://github.com/wildfly/wildfly-openssl">wildlfy-openssl</a> * library.
<p> * Default_JSSE is not truly the the default JSSE implementation because * the GCM cipher is disabled when running on Java 8. However, the name * was not changed in order to preserve backwards compatibility. Instead, * a new mode called Default_JSSE_with_GCM delegates to the default JSSE * implementation with no changes to the list of enabled ciphers. * </p>
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
DelegationTokenIdentifier.java
* A delegation token identifier that is specific to HDFS.
A delegation token identifier that is specific to HDFS.
A delegation token identifier that is specific to HDFS.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
DeleteApplicationHomeSubClusterRequest.java
* The request to <code>Federation state store</code> to delete the mapping of * home subcluster of a submitted application.
The request to <code>Federation state store</code> to delete the mapping of * home subcluster of a submitted application.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
DFSConfigKeys.java
* This class contains constants for configuration keys and default values * used in hdfs.
This class contains constants for configuration keys and default values used in hdfs.
null
This class contains constants for configuration keys and default values used in hdfs.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
DfsServlet.java
* A base class for the servlets in DFS.
A base class for the servlets in DFS
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
DiskBalancerCluster.java
* DiskBalancerCluster represents the nodes that we are working against. * <p> * Please Note : * Semantics of inclusionList and exclusionLists. * <p> * If a non-empty inclusionList is specified then the diskBalancer assumes that * the user is only interested in processing that list of nodes. This node list * is checked against the exclusionList and only the nodes in inclusionList but * not in exclusionList is processed. * <p> * if inclusionList is empty, then we assume that all live nodes in the nodes is * to be processed by diskBalancer. In that case diskBalancer will avoid any * nodes specified in the exclusionList but will process all nodes in the * cluster. * <p> * In other words, an empty inclusionList is means all the nodes otherwise * only a given list is processed and ExclusionList is always honored.
DiskBalancerCluster represents the nodes that we are working against.
Please Note : * Semantics of inclusionList and exclusionLists. * <p> * If a non-empty inclusionList is specified then the diskBalancer assumes that * the user is only interested in processing that list of nodes. This node list * is checked against the exclusionList and only the nodes in inclusionList but * not in exclusionList is processed. * <p> * if inclusionList is empty, then we assume that all live nodes in the nodes is * to be processed by diskBalancer. In that case diskBalancer will avoid any * nodes specified in the exclusionList but will process all nodes in the * cluster. * <p> * In other words, an empty inclusionList is means all the nodes otherwise * only a given list is processed and ExclusionList is always honored.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
DistributedSchedulingAllocateRequest.java
* Object used by the Application Master when distributed scheduling is enabled, * in order to forward the {@link AllocateRequest} for GUARANTEED containers to * the Resource Manager, and to notify the Resource Manager about the allocation * of OPPORTUNISTIC containers through the Distributed Scheduler.
* Object used by the Application Master when distributed scheduling is enabled,
in order to forward the {@link AllocateRequest} for GUARANTEED containers to * the Resource Manager, and to notify the Resource Manager about the allocation * of OPPORTUNISTIC containers through the Distributed Scheduler.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
DockerKillCommand.java
* Encapsulates the docker kill command and its command line arguments.
Encapsulates the docker kill command and its command line arguments.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
EditLogTailer.java
* EditLogTailer represents a thread which periodically reads from edits * journals and applies the transactions contained within to a given * FSNamesystem. | * The thread which does the actual work of tailing edits journals and * applying the transactions to the FSNS. | * Manage the 'active namenode proxy'. This cannot just be the a single proxy since we could * failover across a number of NameNodes, rather than just between an active and a standby. * <p> * We - lazily - get a proxy to one of the configured namenodes and attempt to make the request * against it. If it doesn't succeed, either because the proxy failed to be created or the request * failed, we try the next NN in the list. We try this up to the configuration maximum number of * retries before throwing up our hands. A working proxy is retained across attempts since we * expect the active NameNode to switch rarely. * <p> * This mechanism is <b>very bad</b> for cases where we care about being <i>fast</i>; it just * blindly goes and tries namenodes.
EditLogTailer represents a thread which periodically reads from edits * journals and applies the transactions contained within to a given * FSNamesystem.
| * The thread which does the actual work of tailing edits journals and * applying the transactions to the FSNS. |
* Manage the 'active namenode proxy'. This cannot just be the a single proxy since we could * failover across a number of NameNodes, rather than just between an active and a standby. * <p> * We - lazily - get a proxy to one of the configured namenodes and attempt to make the request * against it. If it doesn't succeed, either because the proxy failed to be created or the request * failed, we try the next NN in the list. We try this up to the configuration maximum number of * retries before throwing up our hands. A working proxy is retained across attempts since we * expect the active NameNode to switch rarely. * <p> * This mechanism is <b>very bad</b> for cases where we care about being <i>fast</i>; it just * blindly goes and tries namenodes.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
ErasureCodingPolicyManager.java
* This manages erasure coding policies predefined and activated in the system. * It loads customized policies and syncs with persisted ones in * NameNode image. * * This class is instantiated by the FSNamesystem.
This manages erasure coding policies predefined and activated in the system
It loads customized policies and syncs with persisted ones in * NameNode image.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
EventWatcher.java
* Event watcher the (re)send a message after timeout. * <p> * Event watcher will send the tracked payload/event after a timeout period * unless a confirmation from the original event (completion event) is arrived. * * @param <TIMEOUT_PAYLOAD> The type of the events which are tracked. * @param <COMPLETION_PAYLOAD> The type of event which could cancel the * tracking.
* Event watcher the (re)send a message after timeout.
null
Event watcher will send the tracked payload/event after a timeout period * unless a confirmation from the original event (completion event) is arrived.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
Expression.java
* Interface describing an expression to be used in the * {@link org.apache.hadoop.fs.shell.find.Find} command.
Interface describing an expression to be used in the * {@link org.apache.hadoop.fs.shell.find.Find} command.
null
to be used in the * {@link org.apache.hadoop.fs.shell.find.Find} command.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
FairOrderingPolicy.java
* An OrderingPolicy which orders SchedulableEntities for fairness (see * FairScheduler * FairSharePolicy), generally, processes with lesser usage are lesser. If * sizedBasedWeight is set to true then an application with high demand * may be prioritized ahead of an application with less usage. This * is to offset the tendency to favor small apps, which could result in * starvation for large apps if many small ones enter and leave the queue * continuously (optional, default false)
* An OrderingPolicy which orders SchedulableEntities for fairness (see * FairScheduler * FairSharePolicy), generally, processes with lesser usage are lesser
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
FederationPolicyException.java
* Generic policy exception.
Generic policy exception
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
FederationProtocolPBTranslator.java
* Helper class for setting/getting data elements in an object backed by a * protobuf implementation.
Helper class for setting/getting data elements in an object backed by a * protobuf implementation.
null
for setting/getting data elements in an object backed by a * protobuf implementation.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
FederationStateStoreInvalidInputException.java
* Exception thrown by the {@code FederationMembershipStateStoreInputValidator}, * {@code FederationApplicationHomeSubClusterStoreInputValidator}, * {@code FederationPolicyStoreInputValidator} if the input is invalid. *
Exception thrown by the {@code FederationMembershipStateStoreInputValidator},
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
FileSystemApplicationHistoryStore.java
* File system implementation of {@link ApplicationHistoryStore}. In this * implementation, one application will have just one file in the file system, * which contains all the history data of one application, and its attempts and * containers. {@link #applicationStarted(ApplicationStartData)} is supposed to * be invoked first when writing any history data of one application and it will * open a file, while {@link #applicationFinished(ApplicationFinishData)} is * supposed to be last writing operation and will close the file.
* File system implementation of {@link ApplicationHistoryStore}.
In this * implementation, one application will have just one file in the file system, * which contains all the history data of one application, and its attempts and * containers. {@link #applicationStarted(ApplicationStartData)} is supposed to * be invoked first when writing any history data of one application and it will * open a file, while {@link #applicationFinished(ApplicationFinishData)} is * supposed to be last writing operation and will close the file.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
FsConstants.java
* FileSystem related constants.
FileSystem related constants
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
GetApplicationHomeSubClusterResponsePBImpl.java
* Protocol buffer based implementation of * {@link GetApplicationHomeSubClusterResponse}.
Protocol buffer based implementation of * {@link GetApplicationHomeSubClusterResponse}.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
GetNamespaceInfoResponse.java
* API response for listing HDFS namespaces present in the state store.
API response for listing HDFS namespaces present in the state store.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
GetNodesToAttributesResponse.java
* <p> * The response sent by the <code>ResourceManager</code> to a client requesting * nodes to attributes mapping. * </p> * * @see ApplicationClientProtocol#getNodesToAttributes * (GetNodesToAttributesRequest)
The response sent by the <code>ResourceManager</code> to a client requesting * nodes to attributes mapping.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
GetSafeModeRequestPBImpl.java
* Protobuf implementation of the state store API object * GetSafeModeRequest.
Protobuf implementation of the state store API object * GetSafeModeRequest.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
GetSubClusterPolicyConfigurationRequestPBImpl.java
* Protocol buffer based implementation of * {@link GetSubClusterPolicyConfigurationRequest}.
Protocol buffer based implementation of * {@link GetSubClusterPolicyConfigurationRequest}.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
HadoopIllegalArgumentException.java
* Indicates that a method has been passed illegal or invalid argument. This * exception is thrown instead of IllegalArgumentException to differentiate the * exception thrown in Hadoop implementation from the one thrown in JDK.
Indicates that a method has been passed illegal or invalid argument.
This * exception is thrown instead of IllegalArgumentException to differentiate the * exception thrown in Hadoop implementation from the one thrown in JDK.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
HashResolver.java
* Order the destinations based on consistent hashing.
Order the destinations based on consistent hashing.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
HttpHeaderConfigurations.java
* Responsible to keep all abfs http headers here.
null
Responsible to keep all abfs http headers here.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
IDataLoader.java
* an IDataLoader loads data on demand
an IDataLoader loads data on demand
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
InconsistentS3ClientFactory.java
* S3 Client factory used for testing with eventual consistency fault injection. * This client is for testing <i>only</i>; it is in the production * {@code hadoop-aws} module to enable integration tests to use this * just by editing the Hadoop configuration used to bring up the client.
S3 Client factory used for testing with eventual consistency fault injection.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
InfoKeyHandler.java
* Executes Info Object.
Executes Info Object
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
InvalidContainerRequestException.java
* Thrown when an arguments are combined to construct a * <code>AMRMClient.ContainerRequest</code> in an invalid way.
* Thrown when an arguments are combined to construct a * <code>AMRMClient.ContainerRequest</code> in an invalid way.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
ITestAbfsReadWriteAndSeek.java
* Test read, write and seek. * Uses package-private methods in AbfsConfiguration, which is why it is in * this package.
Test read, write and seek.
null
Uses package-private methods in AbfsConfiguration, which is why it is in * this package.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
ITestAzureNativeContractSeek.java
* Contract test.
Contract test
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
ITestCommitOperations.java
* Test the low-level binding of the S3A FS to the magic commit mechanism, * and handling of the commit operations. * This is done with an inconsistent client.
* Test the low-level binding of the S3A FS to the magic commit mechanism, * and handling of the commit operations.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
ITestListPerformance.java
* Test list performance.
Test list performance
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
ITestS3Select.java
* Test the S3 Select feature with some basic SQL Commands. * Executed if the destination store declares its support for the feature.
Test the S3 Select feature with some basic SQL Commands.
Executed if the destination store declares its support for the feature.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
ITestS3SelectCLI.java
* Test the S3 Select CLI through some operations against landsat * and files generated from it.
Test the S3 Select CLI through some operations against landsat * and files generated from it.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
KerberosDelegationTokenAuthenticator.java
* The <code>KerberosDelegationTokenAuthenticator</code> provides support for * Kerberos SPNEGO authentication mechanism and support for Hadoop Delegation * Token operations. * <p> * It falls back to the {@link PseudoDelegationTokenAuthenticator} if the HTTP * endpoint does not trigger a SPNEGO authentication
The <code>KerberosDelegationTokenAuthenticator</code> provides support for * Kerberos SPNEGO authentication mechanism and support for Hadoop Delegation * Token operations.
It falls back to the {@link PseudoDelegationTokenAuthenticator} if the HTTP * endpoint does not trigger a SPNEGO authentication
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
LocalizationStatusPBImpl.java
* PB Impl of {@link LocalizationStatus}.
PB Impl of {@link LocalizationStatus}.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
LocatedFileStatus.java
* This class defines a FileStatus that includes a file's block locations.
This class defines a FileStatus that includes a file's block locations
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
LoggedTask.java
* A {@link LoggedTask} represents a [hadoop] task that is part of a hadoop job. * It knows about the [pssibly empty] sequence of attempts, its I/O footprint, * and its runtime. * * All of the public methods are simply accessors for the instance variables we * want to write out in the JSON files. *
A {@link LoggedTask} represents a [hadoop] task that is part of a hadoop job.
It knows about the [pssibly empty] sequence of attempts, its I/O footprint, * and its runtime.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
LogParserUtil.java
* Common utility functions for {@link LogParser}.
Common utility functions for {@link LogParser}
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
LogWebService.java
* Support only ATSv2 client only.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
LRUCacheHashMap.java
* LRU cache with a configurable maximum cache size and access order.
LRU cache with a configurable maximum cache size and access order.
LRU cache with a configurable maximum cache size and access order.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
MapContext.java
* The context that is given to the {@link Mapper}. * @param <KEYIN> the key input type to the Mapper * @param <VALUEIN> the value input type to the Mapper * @param <KEYOUT> the key output type from the Mapper * @param <VALUEOUT> the value output type from the Mapper
The context that is given to the {@link Mapper}.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
MetaBlockAlreadyExists.java
* Exception - Meta Block with the same name already exists.
Exception - Meta Block with the same name already exists.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
MetricsCache.java
* A metrics cache for sinks that don't support sparse updates. | * Cached record
A metrics cache for sinks that don't support sparse updates. Context of the Queues in Scheduler.
Cached record
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
NativeBatchProcessor.java
* used to create channel, transfer data and command between Java and native
null
null
used to create channel, transfer data and command between Java and native
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
NativeSingleLineParser.java
* This sample parser will parse the sample log and extract the resource * skyline. * <p> The expected log format is: NormalizedJobName NumInstances SubmitTime * StartTime EndTime JobInstanceName memUsage coreUsage
This sample parser will parse the sample log and extract the resource * skyline.
<p> The expected log format is: NormalizedJobName NumInstances SubmitTime * StartTime EndTime JobInstanceName memUsage coreUsage
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
Nfs3Metrics.java
* This class is for maintaining the various NFS gateway activity statistics and * publishing them through the metrics interfaces.
This class is for maintaining the various NFS gateway activity statistics and * publishing them through the metrics interfaces.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
Nfs3Status.java
* Success or error status is reported in NFS3 responses.
* Success or error status is reported in NFS3 responses.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
Node2ObjectsMap.java
* This data structure maintains the list of containers that is on a datanode. * This information is built from the DN container reports.
This data structure maintains the list of containers that is on a datanode.
This information is built from the DN container reports.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
NodeUpdateType.java
* <p>Taxonomy of the <code>NodeState</code> that a * <code>Node</code> might transition into.</p> *
<p>Taxonomy of the <code>NodeState</code> that a * <code>Node</code> might transition into.</p>
<p>Taxonomy of the <code>NodeState</code> that a * <code>Node</code> might transition into.</p>
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
NullOutputFormat.java
* Consume all outputs and put them in /dev/null.
Consume all outputs and put them in /dev/null.
Consume all outputs and put them in /dev/null.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
OMNodeDetails.java
* This class stores OM node details. | * Builder class for OMNodeDetails.
* This class stores OM node details. | * Builder class for OMNodeDetails.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
OpportunisticContainersStatusPBImpl.java
* Protocol Buffer implementation of OpportunisticContainersStatus.
Protocol Buffer implementation of OpportunisticContainersStatus
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
OzoneObj.java
* Class representing an unique ozone object. * | * Ozone Objects supported for ACL. | * Ozone Objects supported for ACL.
Class representing an unique ozone object
Ozone Objects supported for ACL.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
Parser.java
A class for parsing outputs
A class for parsing outputs
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
PartialOutputCommitter.java
* Interface for an {@link org.apache.hadoop.mapreduce.OutputCommitter} * implementing partial commit of task output, as during preemption.
* Interface for an {@link org.apache.hadoop.mapreduce.OutputCommitter} * implementing partial commit of task output, as during preemption.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
PartitionResourcesInfo.java
* This class represents queue/user resource usage info for a given partition
This class represents queue/user resource usage info for a given partition
null
This class represents queue/user resource usage info for a given partition
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
PlanningQuotaException.java
* This exception is thrown if the user quota is exceed while accepting or * updating a reservation.
This exception is thrown if the user quota is exceed while accepting or * updating a reservation.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
ProcessIdFileReader.java
* Helper functionality to read the pid from a file.
Helper functionality to read the pid from a file.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
QuasiMonteCarlo.java
* A map/reduce program that estimates the value of Pi * using a quasi-Monte Carlo (qMC) method. * Arbitrary integrals can be approximated numerically by qMC methods. * In this example, * we use a qMC method to approximate the integral $I = \int_S f(x) dx$, * where $S=[0,1)^2$ is a unit square, * $x=(x_1,x_2)$ is a 2-dimensional point, * and $f$ is a function describing the inscribed circle of the square $S$, * $f(x)=1$ if $(2x_1-1)^2+(2x_2-1)^2 &lt;= 1$ and $f(x)=0$, otherwise. * It is easy to see that Pi is equal to $4I$. * So an approximation of Pi is obtained once $I$ is evaluated numerically. * * There are better methods for computing Pi. * We emphasize numerical approximation of arbitrary integrals in this example. * For computing many digits of Pi, consider using bbp. * * The implementation is discussed below. * * Mapper: * Generate points in a unit square * and then count points inside/outside of the inscribed circle of the square. * * Reducer: * Accumulate points inside/outside results from the mappers. * * Let numTotal = numInside + numOutside. * The fraction numInside/numTotal is a rational approximation of * the value (Area of the circle)/(Area of the square) = $I$, * where the area of the inscribed circle is Pi/4 * and the area of unit square is 1. * Finally, the estimated value of Pi is 4(numInside/numTotal). | 2-dimensional Halton sequence {H(i)}, * where H(i) is a 2-dimensional point and i >= 1 is the index. * Halton sequence is used to generate sample points for Pi estimation. | * Mapper class for Pi estimation. * Generate points in a unit square * and then count points inside/outside of the inscribed circle of the square. | * Reducer class for Pi estimation. * Accumulate points inside/outside results from the mappers.
* A map/reduce program that estimates the value of Pi.
using a quasi-Monte Carlo (qMC) method. Arbitrary integrals can be approximated numerically by qMC methods. * There are better methods for computing Pi. * We emphasize numerical approximation of arbitrary integrals in this example. * For computing many digits of Pi, consider using bbp. * * The implementation is discussed below. * * Mapper: * Generate points in a unit square * and then count points inside/outside of the inscribed circle of the square. * * Reducer: * Accumulate points inside/outside results from the mappers. * Mapper class for Pi estimation. * Generate points in a unit square * and then count points inside/outside of the inscribed circle of the square. | * Reducer class for Pi estimation. * Accumulate points inside/outside results from the mappers.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
Query.java
* Check if a record matches a query. The query is usually a partial record. * * @param <T> Type of the record to query.
Check if a record matches a query.
The query is usually a partial record.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
QueueName.java
* Represents a queue name.
Represents a queue name.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
RandomKeyGenerator.java
* Data generator tool to generate as much keys as possible. | * Wrapper to hold ozone keyValidate entry. | * Validates the write done in ozone cluster.
Data generator tool to generate as much keys as possible.
Wrapper to hold ozone keyValidate entry. | * Validates the write done in ozone cluster.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
ReencryptionUpdater.java
* Class for finalizing re-encrypt EDEK operations, by updating file xattrs with * edeks returned from reencryption. * <p> * The tasks are submitted by ReencryptionHandler. * <p> * It is assumed only 1 Updater will be running, since updating file xattrs * requires namespace write lock, and performance gain from multi-threading * is limited. | * Class to track re-encryption submissions of a single zone. It contains * all the submitted futures, and statistics about how far the futures are * processed. | * Class representing the task for one batch of a re-encryption command. It * also contains statistics about how far this single batch has been executed. | * Class that encapsulates re-encryption details of a file.
Class for finalizing re-encrypt EDEK operations. Class representing the task for one batch of a re-encryption command. It * also contains statistics about how far this single batch has been executed. Class that encapsulates re-encryption details of a file.
by updating file xattrs with * edeks returned from reencryption. * <p> * The tasks are submitted by ReencryptionHandler. * <p>. It is assumed only 1 Updater will be running, since updating file xattrs * requires namespace write lock, and performance gain from multi-threading * is limited. | * Class to track re-encryption submissions of a single zone. It contains * all the submitted futures, and statistics about how far the futures are * processed. It contains the * file inode, stores the initial edek of the file, and the new edek * after re-encryption. * <p> * Assumptions are the object initialization happens when dir lock is held, * and inode is valid and is encrypted during initialization. * <p> * Namespace changes may happen during re-encryption, and if inode is changed * the re-encryption is skipped.
Class for finalizing re-encrypt EDEK operations
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
RegistryInternalConstants.java
* Internal constants for the registry. * * These are the things which aren't visible to users. *
Internal constants for the registry.
null
These are the things which aren't visible to users.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
RegistryOperations.java
* Registry Operations
Registry Operations
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
ReInitializeContainerRequestPBImpl.java
CHECKSTYLE:OFF
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
ResourceBlacklistRequest.java
* {@link ResourceBlacklistRequest} encapsulates the list of resource-names * which should be added or removed from the <em>blacklist</em> of resources * for the application. * * @see ResourceRequest * @see ApplicationMasterProtocol#allocate(org.apache.hadoop.yarn.api.protocolrecords.AllocateRequest)
* {@link ResourceBlacklistRequest} encapsulates the list of resource-names * which should be added or removed from the <em>blacklist</em> of resources * for the application.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
ResourceRequestsJsonVerifications.java
* Performs value verifications on * {@link org.apache.hadoop.yarn.server.resourcemanager.webapp.dao.ResourceRequestInfo} * objects against the values of {@link ResourceRequest}. With the help of the * {@link Builder}, users can also make verifications of the custom resource * types and its values. | * Builder class for {@link ResourceRequestsJsonVerifications}.
* Performs value verifications on * {@link org.apache.hadoop.yarn.server.resourcemanager.webapp.dao.ResourceRequestInfo} * objects against the values of {@link ResourceRequest}. | * Builder class for {@link ResourceRequestsJsonVerifications}.
With the help of the * {@link Builder}, users can also make verifications of the custom resource * types and its values.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
RetriableDirectoryCreateCommand.java
* This class extends Retriable command to implement the creation of directories * with retries on failure.
This class extends Retriable command to implement the creation of directories * with retries on failure.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
RMAdminRequestInterceptor.java
* Defines the contract to be implemented by the request intercepter classes, * that can be used to intercept and inspect messages sent from the client to * the resource manager.
Defines the contract to be implemented by the request intercepter classes
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
RSLegacyRawErasureCoderFactory.java
* A raw coder factory for the legacy raw Reed-Solomon coder in Java.
A raw coder factory for the legacy raw Reed-Solomon coder in Java.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
SafeModeException.java
* This exception is thrown when the name node is in safe mode. * Client cannot modified namespace until the safe mode is off. *
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
SchedulerQueueManager.java
* * Context of the Queues in Scheduler. *
Context of the Queues in Scheduler.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
1
SequenceFileRecordReader.java
* An {@link RecordReader} for {@link SequenceFile}s.
* An {@link RecordReader} for {@link SequenceFile}s.
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
End of preview.

Dataset Card for Code Comment Classification

Dataset Summary

The dataset contains class comments extracted from various big and diverse open-source projects of three programming languages Java, Smalltalk, and Python.

Supported Tasks and Leaderboards

Single-label text classification and Multi-label text classification

Languages

Java, Python, Smalltalk

Dataset Structure

Data Instances

{
  "class" : "Absy.java",
  "comment":"* Azure Blob File System implementation of AbstractFileSystem. * This impl delegates to the old FileSystem",
  "summary":"Azure Blob File System implementation of AbstractFileSystem.",
  "expand":"This impl delegates to the old FileSystem",
  "rational":"",
  "deprecation":"",
  "usage":"",
  "exception":"",
  "todo":"",
  "incomplete":"",
  "commentedcode":"",
  "directive":"",
  "formatter":"",
  "license":"",
  "ownership":"",
  "pointer":"",
  "autogenerated":"",
  "noise":"",
  "warning":"",
  "recommendation":"",
  "precondition":"",
  "codingGuidelines":"",
  "extension":"",
  "subclassexplnation":"",
  "observation":"",
}

Data Fields

class: name of the class with the language extension.

comment: class comment of the class

categories: a category that sentence is classified to. It indicated a particular type of information.

Data Splits

10-fold cross validation

Dataset Creation

Curation Rationale

To identify the infomation embedded in the class comments across various projects and programming languages.

Source Data

Initial Data Collection and Normalization

It contains the dataset extracted from various open-source projects of three programming languages Java, Smalltalk, and Python.

  • Java

    Each file contains all the extracted class comments from one project. We have a total of six java projects. We chose a sample of 350 comments from all these files for our experiment.

    • Eclipse.csv - Extracted class comments from the Eclipse project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub Eclipse.

    • Guava.csv - Extracted class comments from the Guava project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub Guava.

    • Guice.csv - Extracted class comments from the Guice project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub Guice.

    • Hadoop.csv - Extracted class comments from the Hadoop project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub Apache Hadoop

    • Spark.csv - Extracted class comments from the Apache Spark project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub Apache Spark

    • Vaadin.csv - Extracted class comments from the Vaadin project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub Vaadin

    • Parser_Details.md - Details of the parser used to parse class comments of Java Projects

  • Smalltalk/

        Each file contains all the extracted class comments from one project. We have a total of seven Pharo projects. We chose a sample of 350 comments from all these files for our experiment.
    
    • GToolkit.csv - Extracted class comments from the GToolkit project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo.

    • Moose.csv - Extracted class comments from the Moose project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo.

    • PetitParser.csv - Extracted class comments from the PetitParser project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo.

    • Pillar.csv - Extracted class comments from the Pillar project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo.

    • PolyMath.csv - Extracted class comments from the PolyMath project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo.

    • Roassal2.csv -Extracted class comments from the Roassal2 project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo.

    • Seaside.csv - Extracted class comments from the Seaside project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo.

    • Parser_Details.md - Details of the parser used to parse class comments of Pharo Projects

  • Python/

        Each file contains all the extracted class comments from one project. We have a total of seven Python projects. We chose a sample of 350 comments from all these files for our experiment.
    
    • Django.csv - Extracted class comments from the Django project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub Django

    • IPython.csv - Extracted class comments from the Ipython project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHubIPython

    • Mailpile.csv - Extracted class comments from the Mailpile project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub Mailpile

    • Pandas.csv - Extracted class comments from the Pandas project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub pandas

    • Pipenv.csv - Extracted class comments from the Pipenv project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub Pipenv

    • Pytorch.csv - Extracted class comments from the Pytorch project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub PyTorch

    • Requests.csv - Extracted class comments from the Requests project. The version of the project referred to extract class comments is available as Raw Dataset on Zenodo. More detail about the project is available on GitHub Requests

    • Parser_Details.md - Details of the parser used to parse class comments of Python Projects

Annotations

Annotation process

Four evaluators (all authors of this paper (https://doi.org/10.1016/j.jss.2021.111047)), each having at least four years of programming experience, participated in the annonation process. We partitioned Java, Python, and Smalltalk comments equally among all evaluators based on the distribution of the language's dataset to ensure the inclusion of comments from all projects and diversified lengths. Each classification is reviewed by three evaluators. The details are given in the paper Rani et al., JSS, 2021

Who are the annotators?

Rani et al., JSS, 2021

Personal and Sensitive Information

Author information embedded in the text

Additional Information

Dataset Curators

[Pooja Rani, Ivan, Manuel]

Licensing Information

[license: cc-by-nc-sa-4.0]

Citation Information

@article{RANI2021111047,
title = {How to identify class comment types? A multi-language approach for class comment classification},
journal = {Journal of Systems and Software},
volume = {181},
pages = {111047},
year = {2021},
issn = {0164-1212},
doi = {https://doi.org/10.1016/j.jss.2021.111047},
url = {https://www.sciencedirect.com/science/article/pii/S0164121221001448},
author = {Pooja Rani and Sebastiano Panichella and Manuel Leuenberger and Andrea {Di Sorbo} and Oscar Nierstrasz},
keywords = {Natural language processing technique, Code comment analysis, Software documentation}
}
Downloads last month
53