text
stringlengths 307
2.02k
| inputs
dict | prediction
null | prediction_agent
null | annotation
stringclasses 2
values | annotation_agent
stringclasses 1
value | vectors
null | multi_label
bool 1
class | explanation
null | id
stringlengths 36
36
| metadata
null | status
stringclasses 2
values | metrics
dict | label
class label 2
classes |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
TITLE: WikiLingua: A New Benchmark Dataset for Cross-Lingual Abstractive Summarization
ABSTRACT: We introduce WikiLingua, a large-scale, multilingual dataset for the
evaluation of crosslingual abstractive summarization systems. We extract
article and summary pairs in 18 languages from WikiHow, a high quality,
collaborative resource of how-to guides on a diverse set of topics written by
human authors. We create gold-standard article-summary alignments across
languages by aligning the images that are used to describe each how-to step in
an article. As a set of baselines for further studies, we evaluate the
performance of existing cross-lingual abstractive summarization methods on our
dataset. We further propose a method for direct crosslingual summarization
(i.e., without requiring translation at inference time) by leveraging synthetic
data and Neural Machine Translation as a pre-training step. Our method
significantly outperforms the baseline approaches, while being more cost
efficient during inference. | {
"abstract": "We introduce WikiLingua, a large-scale, multilingual dataset for the\nevaluation of crosslingual abstractive summarization systems. We extract\narticle and summary pairs in 18 languages from WikiHow, a high quality,\ncollaborative resource of how-to guides on a diverse set of topics written by\nhuman authors. We create gold-standard article-summary alignments across\nlanguages by aligning the images that are used to describe each how-to step in\nan article. As a set of baselines for further studies, we evaluate the\nperformance of existing cross-lingual abstractive summarization methods on our\ndataset. We further propose a method for direct crosslingual summarization\n(i.e., without requiring translation at inference time) by leveraging synthetic\ndata and Neural Machine Translation as a pre-training step. Our method\nsignificantly outperforms the baseline approaches, while being more cost\nefficient during inference.",
"title": "WikiLingua: A New Benchmark Dataset for Cross-Lingual Abstractive Summarization",
"url": "http://arxiv.org/abs/2010.03093v1"
} | null | null | new_dataset | admin | null | false | null | 2b06bf83-4565-40c7-bd94-55d327c90489 | null | Validated | {
"text_length": 1034
} | 0new_dataset
|
TITLE: Machine Learning Workflow to Explain Black-box Models for Early Alzheimer's Disease Classification Evaluated for Multiple Datasets
ABSTRACT: Purpose: Hard-to-interpret Black-box Machine Learning (ML) were often used
for early Alzheimer's Disease (AD) detection.
Methods: To interpret eXtreme Gradient Boosting (XGBoost), Random Forest
(RF), and Support Vector Machine (SVM) black-box models a workflow based on
Shapley values was developed. All models were trained on the Alzheimer's
Disease Neuroimaging Initiative (ADNI) dataset and evaluated for an independent
ADNI test set, as well as the external Australian Imaging and Lifestyle
flagship study of Ageing (AIBL), and Open Access Series of Imaging Studies
(OASIS) datasets. Shapley values were compared to intuitively interpretable
Decision Trees (DTs), and Logistic Regression (LR), as well as natural and
permutation feature importances. To avoid the reduction of the explanation
validity caused by correlated features, forward selection and aspect
consolidation were implemented.
Results: Some black-box models outperformed DTs and LR. The forward-selected
features correspond to brain areas previously associated with AD. Shapley
values identified biologically plausible associations with moderate to strong
correlations with feature importances. The most important RF features to
predict AD conversion were the volume of the amygdalae, and a cognitive test
score. Good cognitive test performances and large brain volumes decreased the
AD risk. The models trained using cognitive test scores significantly
outperformed brain volumetric models ($p<0.05$). Cognitive Normal (CN) vs. AD
models were successfully transferred to external datasets.
Conclusion: In comparison to previous work, improved performances for ADNI
and AIBL were achieved for CN vs. Mild Cognitive Impairment (MCI)
classification using brain volumes. The Shapley values and the feature
importances showed moderate to strong correlations. | {
"abstract": "Purpose: Hard-to-interpret Black-box Machine Learning (ML) were often used\nfor early Alzheimer's Disease (AD) detection.\n Methods: To interpret eXtreme Gradient Boosting (XGBoost), Random Forest\n(RF), and Support Vector Machine (SVM) black-box models a workflow based on\nShapley values was developed. All models were trained on the Alzheimer's\nDisease Neuroimaging Initiative (ADNI) dataset and evaluated for an independent\nADNI test set, as well as the external Australian Imaging and Lifestyle\nflagship study of Ageing (AIBL), and Open Access Series of Imaging Studies\n(OASIS) datasets. Shapley values were compared to intuitively interpretable\nDecision Trees (DTs), and Logistic Regression (LR), as well as natural and\npermutation feature importances. To avoid the reduction of the explanation\nvalidity caused by correlated features, forward selection and aspect\nconsolidation were implemented.\n Results: Some black-box models outperformed DTs and LR. The forward-selected\nfeatures correspond to brain areas previously associated with AD. Shapley\nvalues identified biologically plausible associations with moderate to strong\ncorrelations with feature importances. The most important RF features to\npredict AD conversion were the volume of the amygdalae, and a cognitive test\nscore. Good cognitive test performances and large brain volumes decreased the\nAD risk. The models trained using cognitive test scores significantly\noutperformed brain volumetric models ($p<0.05$). Cognitive Normal (CN) vs. AD\nmodels were successfully transferred to external datasets.\n Conclusion: In comparison to previous work, improved performances for ADNI\nand AIBL were achieved for CN vs. Mild Cognitive Impairment (MCI)\nclassification using brain volumes. The Shapley values and the feature\nimportances showed moderate to strong correlations.",
"title": "Machine Learning Workflow to Explain Black-box Models for Early Alzheimer's Disease Classification Evaluated for Multiple Datasets",
"url": "http://arxiv.org/abs/2205.05907v2"
} | null | null | no_new_dataset | admin | null | false | null | 36f5d66c-94a0-4d1c-88c0-f8cd4806b0b1 | null | Validated | {
"text_length": 1995
} | 1no_new_dataset
|
TITLE: A Case Study on the Classification of Lost Circulation Events During Drilling using Machine Learning Techniques on an Imbalanced Large Dataset
ABSTRACT: This study presents machine learning models that forecast and categorize lost
circulation severity preemptively using a large class imbalanced drilling
dataset. We demonstrate reproducible core techniques involved in tackling a
large drilling engineering challenge utilizing easily interpretable machine
learning approaches.
We utilized a 65,000+ records data with class imbalance problem from Azadegan
oilfield formations in Iran. Eleven of the dataset's seventeen parameters are
chosen to be used in the classification of five lost circulation events. To
generate classification models, we used six basic machine learning algorithms
and four ensemble learning methods. Linear Discriminant Analysis (LDA),
Logistic Regression (LR), Support Vector Machines (SVM), Classification and
Regression Trees (CART), k-Nearest Neighbors (KNN), and Gaussian Naive Bayes
(GNB) are the six fundamental techniques. We also used bagging and boosting
ensemble learning techniques in the investigation of solutions for improved
predicting performance. The performance of these algorithms is measured using
four metrics: accuracy, precision, recall, and F1-score. The F1-score weighted
to represent the data imbalance is chosen as the preferred evaluation
criterion.
The CART model was found to be the best in class for identifying drilling
fluid circulation loss events with an average weighted F1-score of 0.9904 and
standard deviation of 0.0015. Upon application of ensemble learning techniques,
a Random Forest ensemble of decision trees showed the best predictive
performance. It identified and classified lost circulation events with a
perfect weighted F1-score of 1.0. Using Permutation Feature Importance (PFI),
the measured depth was found to be the most influential factor in accurately
recognizing lost circulation events while drilling. | {
"abstract": "This study presents machine learning models that forecast and categorize lost\ncirculation severity preemptively using a large class imbalanced drilling\ndataset. We demonstrate reproducible core techniques involved in tackling a\nlarge drilling engineering challenge utilizing easily interpretable machine\nlearning approaches.\n We utilized a 65,000+ records data with class imbalance problem from Azadegan\noilfield formations in Iran. Eleven of the dataset's seventeen parameters are\nchosen to be used in the classification of five lost circulation events. To\ngenerate classification models, we used six basic machine learning algorithms\nand four ensemble learning methods. Linear Discriminant Analysis (LDA),\nLogistic Regression (LR), Support Vector Machines (SVM), Classification and\nRegression Trees (CART), k-Nearest Neighbors (KNN), and Gaussian Naive Bayes\n(GNB) are the six fundamental techniques. We also used bagging and boosting\nensemble learning techniques in the investigation of solutions for improved\npredicting performance. The performance of these algorithms is measured using\nfour metrics: accuracy, precision, recall, and F1-score. The F1-score weighted\nto represent the data imbalance is chosen as the preferred evaluation\ncriterion.\n The CART model was found to be the best in class for identifying drilling\nfluid circulation loss events with an average weighted F1-score of 0.9904 and\nstandard deviation of 0.0015. Upon application of ensemble learning techniques,\na Random Forest ensemble of decision trees showed the best predictive\nperformance. It identified and classified lost circulation events with a\nperfect weighted F1-score of 1.0. Using Permutation Feature Importance (PFI),\nthe measured depth was found to be the most influential factor in accurately\nrecognizing lost circulation events while drilling.",
"title": "A Case Study on the Classification of Lost Circulation Events During Drilling using Machine Learning Techniques on an Imbalanced Large Dataset",
"url": "http://arxiv.org/abs/2209.01607v2"
} | null | null | no_new_dataset | admin | null | false | null | aa7fbe27-3a54-4e61-9172-64185ae18634 | null | Validated | {
"text_length": 2012
} | 1no_new_dataset
|
TITLE: Presenting a Larger Up-to-date Movie Dataset and Investigating the Effects of Pre-released Attributes on Gross Revenue
ABSTRACT: Movie-making has become one of the most costly and risky endeavors in the
entertainment industry. Continuous change in the preference of the audience
makes it harder to predict what kind of movie will be financially successful at
the box office. So, it is no wonder that cautious, intelligent stakeholders and
large production houses will always want to know the probable revenue that will
be generated by a movie before making an investment. Researchers have been
working on finding an optimal strategy to help investors in making the right
decisions. But the lack of a large, up-to-date dataset makes their work harder.
In this work, we introduce an up-to-date, richer, and larger dataset that we
have prepared by scraping IMDb for researchers and data analysts to work with.
The compiled dataset contains the summery data of 7.5 million titles and detail
information of more than 200K movies. Additionally, we perform different
statistical analysis approaches on our dataset to find out how a movie's
revenue is affected by different pre-released attributes such as budget,
runtime, release month, content rating, genre etc. In our analysis, we have
found that having a star cast/director has a positive impact on generated
revenue. We introduce a novel approach for calculating the star power of a
movie. Based on our analysis we select a set of attributes as features and
train different machine learning algorithms to predict a movie's expected
revenue. Based on generated revenue, we classified the movies in 10 categories
and achieved a one-class-away accuracy rate of almost 60% (bingo accuracy of
30%). All the generated datasets and analysis codes are available online. We
also made the source codes of our scraper bots public, so that researchers
interested in extending this work can easily modify these bots as they need and
prepare their own up-to-date datasets. | {
"abstract": "Movie-making has become one of the most costly and risky endeavors in the\nentertainment industry. Continuous change in the preference of the audience\nmakes it harder to predict what kind of movie will be financially successful at\nthe box office. So, it is no wonder that cautious, intelligent stakeholders and\nlarge production houses will always want to know the probable revenue that will\nbe generated by a movie before making an investment. Researchers have been\nworking on finding an optimal strategy to help investors in making the right\ndecisions. But the lack of a large, up-to-date dataset makes their work harder.\nIn this work, we introduce an up-to-date, richer, and larger dataset that we\nhave prepared by scraping IMDb for researchers and data analysts to work with.\nThe compiled dataset contains the summery data of 7.5 million titles and detail\ninformation of more than 200K movies. Additionally, we perform different\nstatistical analysis approaches on our dataset to find out how a movie's\nrevenue is affected by different pre-released attributes such as budget,\nruntime, release month, content rating, genre etc. In our analysis, we have\nfound that having a star cast/director has a positive impact on generated\nrevenue. We introduce a novel approach for calculating the star power of a\nmovie. Based on our analysis we select a set of attributes as features and\ntrain different machine learning algorithms to predict a movie's expected\nrevenue. Based on generated revenue, we classified the movies in 10 categories\nand achieved a one-class-away accuracy rate of almost 60% (bingo accuracy of\n30%). All the generated datasets and analysis codes are available online. We\nalso made the source codes of our scraper bots public, so that researchers\ninterested in extending this work can easily modify these bots as they need and\nprepare their own up-to-date datasets.",
"title": "Presenting a Larger Up-to-date Movie Dataset and Investigating the Effects of Pre-released Attributes on Gross Revenue",
"url": "http://arxiv.org/abs/2110.07039v2"
} | null | null | new_dataset | admin | null | false | null | 14366fed-8b51-40fd-9cc3-c8ff049dd855 | null | Validated | {
"text_length": 2030
} | 0new_dataset
|
TITLE: Human-Machine Collaboration Approaches to Build a Dialogue Dataset for Hate Speech Countering
ABSTRACT: Fighting online hate speech is a challenge that is usually addressed using
Natural Language Processing via automatic detection and removal of hate
content. Besides this approach, counter narratives have emerged as an effective
tool employed by NGOs to respond to online hate on social media platforms. For
this reason, Natural Language Generation is currently being studied as a way to
automatize counter narrative writing. However, the existing resources necessary
to train NLG models are limited to 2-turn interactions (a hate speech and a
counter narrative as response), while in real life, interactions can consist of
multiple turns. In this paper, we present a hybrid approach for dialogical data
collection, which combines the intervention of human expert annotators over
machine generated dialogues obtained using 19 different configurations. The
result of this work is DIALOCONAN, the first dataset comprising over 3000
fictitious multi-turn dialogues between a hater and an NGO operator, covering 6
targets of hate. | {
"abstract": "Fighting online hate speech is a challenge that is usually addressed using\nNatural Language Processing via automatic detection and removal of hate\ncontent. Besides this approach, counter narratives have emerged as an effective\ntool employed by NGOs to respond to online hate on social media platforms. For\nthis reason, Natural Language Generation is currently being studied as a way to\nautomatize counter narrative writing. However, the existing resources necessary\nto train NLG models are limited to 2-turn interactions (a hate speech and a\ncounter narrative as response), while in real life, interactions can consist of\nmultiple turns. In this paper, we present a hybrid approach for dialogical data\ncollection, which combines the intervention of human expert annotators over\nmachine generated dialogues obtained using 19 different configurations. The\nresult of this work is DIALOCONAN, the first dataset comprising over 3000\nfictitious multi-turn dialogues between a hater and an NGO operator, covering 6\ntargets of hate.",
"title": "Human-Machine Collaboration Approaches to Build a Dialogue Dataset for Hate Speech Countering",
"url": "http://arxiv.org/abs/2211.03433v1"
} | null | null | new_dataset | admin | null | false | null | 7846bbb9-9c43-48f8-8277-1d711ec791c7 | null | Validated | {
"text_length": 1152
} | 0new_dataset
|
TITLE: MultiCoNER: A Large-scale Multilingual dataset for Complex Named Entity Recognition
ABSTRACT: We present MultiCoNER, a large multilingual dataset for Named Entity
Recognition that covers 3 domains (Wiki sentences, questions, and search
queries) across 11 languages, as well as multilingual and code-mixing subsets.
This dataset is designed to represent contemporary challenges in NER, including
low-context scenarios (short and uncased text), syntactically complex entities
like movie titles, and long-tail entity distributions. The 26M token dataset is
compiled from public resources using techniques such as heuristic-based
sentence sampling, template extraction and slotting, and machine translation.
We applied two NER models on our dataset: a baseline XLM-RoBERTa model, and a
state-of-the-art GEMNET model that leverages gazetteers. The baseline achieves
moderate performance (macro-F1=54%), highlighting the difficulty of our data.
GEMNET, which uses gazetteers, improvement significantly (average improvement
of macro-F1=+30%). MultiCoNER poses challenges even for large pre-trained
language models, and we believe that it can help further research in building
robust NER systems. MultiCoNER is publicly available at
https://registry.opendata.aws/multiconer/ and we hope that this resource will
help advance research in various aspects of NER. | {
"abstract": "We present MultiCoNER, a large multilingual dataset for Named Entity\nRecognition that covers 3 domains (Wiki sentences, questions, and search\nqueries) across 11 languages, as well as multilingual and code-mixing subsets.\nThis dataset is designed to represent contemporary challenges in NER, including\nlow-context scenarios (short and uncased text), syntactically complex entities\nlike movie titles, and long-tail entity distributions. The 26M token dataset is\ncompiled from public resources using techniques such as heuristic-based\nsentence sampling, template extraction and slotting, and machine translation.\nWe applied two NER models on our dataset: a baseline XLM-RoBERTa model, and a\nstate-of-the-art GEMNET model that leverages gazetteers. The baseline achieves\nmoderate performance (macro-F1=54%), highlighting the difficulty of our data.\nGEMNET, which uses gazetteers, improvement significantly (average improvement\nof macro-F1=+30%). MultiCoNER poses challenges even for large pre-trained\nlanguage models, and we believe that it can help further research in building\nrobust NER systems. MultiCoNER is publicly available at\nhttps://registry.opendata.aws/multiconer/ and we hope that this resource will\nhelp advance research in various aspects of NER.",
"title": "MultiCoNER: A Large-scale Multilingual dataset for Complex Named Entity Recognition",
"url": "http://arxiv.org/abs/2208.14536v1"
} | null | null | new_dataset | admin | null | false | null | 56aa19a9-b936-46d9-9501-40cba4082c0e | null | Validated | {
"text_length": 1375
} | 0new_dataset
|
TITLE: Healthsheet: Development of a Transparency Artifact for Health Datasets
ABSTRACT: Machine learning (ML) approaches have demonstrated promising results in a
wide range of healthcare applications. Data plays a crucial role in developing
ML-based healthcare systems that directly affect people's lives. Many of the
ethical issues surrounding the use of ML in healthcare stem from structural
inequalities underlying the way we collect, use, and handle data. Developing
guidelines to improve documentation practices regarding the creation, use, and
maintenance of ML healthcare datasets is therefore of critical importance. In
this work, we introduce Healthsheet, a contextualized adaptation of the
original datasheet questionnaire ~\cite{gebru2018datasheets} for
health-specific applications. Through a series of semi-structured interviews,
we adapt the datasheets for healthcare data documentation. As part of the
Healthsheet development process and to understand the obstacles researchers
face in creating datasheets, we worked with three publicly-available healthcare
datasets as our case studies, each with different types of structured data:
Electronic health Records (EHR), clinical trial study data, and
smartphone-based performance outcome measures. Our findings from the
interviewee study and case studies show 1) that datasheets should be
contextualized for healthcare, 2) that despite incentives to adopt
accountability practices such as datasheets, there is a lack of consistency in
the broader use of these practices 3) how the ML for health community views
datasheets and particularly \textit{Healthsheets} as diagnostic tool to surface
the limitations and strength of datasets and 4) the relative importance of
different fields in the datasheet to healthcare concerns. | {
"abstract": "Machine learning (ML) approaches have demonstrated promising results in a\nwide range of healthcare applications. Data plays a crucial role in developing\nML-based healthcare systems that directly affect people's lives. Many of the\nethical issues surrounding the use of ML in healthcare stem from structural\ninequalities underlying the way we collect, use, and handle data. Developing\nguidelines to improve documentation practices regarding the creation, use, and\nmaintenance of ML healthcare datasets is therefore of critical importance. In\nthis work, we introduce Healthsheet, a contextualized adaptation of the\noriginal datasheet questionnaire ~\\cite{gebru2018datasheets} for\nhealth-specific applications. Through a series of semi-structured interviews,\nwe adapt the datasheets for healthcare data documentation. As part of the\nHealthsheet development process and to understand the obstacles researchers\nface in creating datasheets, we worked with three publicly-available healthcare\ndatasets as our case studies, each with different types of structured data:\nElectronic health Records (EHR), clinical trial study data, and\nsmartphone-based performance outcome measures. Our findings from the\ninterviewee study and case studies show 1) that datasheets should be\ncontextualized for healthcare, 2) that despite incentives to adopt\naccountability practices such as datasheets, there is a lack of consistency in\nthe broader use of these practices 3) how the ML for health community views\ndatasheets and particularly \\textit{Healthsheets} as diagnostic tool to surface\nthe limitations and strength of datasets and 4) the relative importance of\ndifferent fields in the datasheet to healthcare concerns.",
"title": "Healthsheet: Development of a Transparency Artifact for Health Datasets",
"url": "http://arxiv.org/abs/2202.13028v1"
} | null | null | no_new_dataset | admin | null | false | null | 11460e75-a3e1-47e4-b475-05f92c8cf564 | null | Validated | {
"text_length": 1803
} | 1no_new_dataset
|
TITLE: CoAP-DoS: An IoT Network Intrusion Dataset
ABSTRACT: The need for secure Internet of Things (IoT) devices is growing as IoT
devices are becoming more integrated into vital networks. Many systems rely on
these devices to remain available and provide reliable service. Denial of
service attacks against IoT devices are a real threat due to the fact these low
power devices are very susceptible to denial-of-service attacks. Machine
learning enabled network intrusion detection systems are effective at
identifying new threats, but they require a large amount of data to work well.
There are many network traffic data sets but very few that focus on IoT network
traffic. Within the IoT network data sets there is a lack of CoAP denial of
service data. We propose a novel data set covering this gap. We develop a new
data set by collecting network traffic from real CoAP denial of service attacks
and compare the data on multiple different machine learning classifiers. We
show that the data set is effective on many classifiers. | {
"abstract": "The need for secure Internet of Things (IoT) devices is growing as IoT\ndevices are becoming more integrated into vital networks. Many systems rely on\nthese devices to remain available and provide reliable service. Denial of\nservice attacks against IoT devices are a real threat due to the fact these low\npower devices are very susceptible to denial-of-service attacks. Machine\nlearning enabled network intrusion detection systems are effective at\nidentifying new threats, but they require a large amount of data to work well.\nThere are many network traffic data sets but very few that focus on IoT network\ntraffic. Within the IoT network data sets there is a lack of CoAP denial of\nservice data. We propose a novel data set covering this gap. We develop a new\ndata set by collecting network traffic from real CoAP denial of service attacks\nand compare the data on multiple different machine learning classifiers. We\nshow that the data set is effective on many classifiers.",
"title": "CoAP-DoS: An IoT Network Intrusion Dataset",
"url": "http://arxiv.org/abs/2206.14341v1"
} | null | null | new_dataset | admin | null | false | null | 96fb32b7-5a67-43e2-bf98-3001963116fe | null | Validated | {
"text_length": 1049
} | 0new_dataset
|
TITLE: Multi-feature Dataset for Windows PE Malware Classification
ABSTRACT: This paper describes a multi-feature dataset for training machine learning
classifiers for detecting malicious Windows Portable Executable (PE) files. The
dataset includes four feature sets from 18,551 binary samples belonging to five
malware families including Spyware, Ransomware, Downloader, Backdoor and
Generic Malware. The feature sets include the list of DLLs and their functions,
values of different fields of PE Header and Sections. First, we explain the
data collection and creation phase and then we explain how did we label the
samples in it using VirusTotal's services. Finally, we explore the dataset to
describe how this dataset can benefit the researchers for static malware
analysis. The dataset is made public in the hope that it will help inspire
machine learning research for malware detection. | {
"abstract": "This paper describes a multi-feature dataset for training machine learning\nclassifiers for detecting malicious Windows Portable Executable (PE) files. The\ndataset includes four feature sets from 18,551 binary samples belonging to five\nmalware families including Spyware, Ransomware, Downloader, Backdoor and\nGeneric Malware. The feature sets include the list of DLLs and their functions,\nvalues of different fields of PE Header and Sections. First, we explain the\ndata collection and creation phase and then we explain how did we label the\nsamples in it using VirusTotal's services. Finally, we explore the dataset to\ndescribe how this dataset can benefit the researchers for static malware\nanalysis. The dataset is made public in the hope that it will help inspire\nmachine learning research for malware detection.",
"title": "Multi-feature Dataset for Windows PE Malware Classification",
"url": "http://arxiv.org/abs/2210.16285v1"
} | null | null | new_dataset | admin | null | false | null | 59ca26fc-a4b8-4f78-94dd-16d31625ab97 | null | Validated | {
"text_length": 908
} | 0new_dataset
|
TITLE: Avast-CTU Public CAPE Dataset
ABSTRACT: There is a limited amount of publicly available data to support research in
malware analysis technology. Particularly, there are virtually no publicly
available datasets generated from rich sandboxes such as Cuckoo/CAPE. The
benefit of using dynamic sandboxes is the realistic simulation of file
execution in the target machine and obtaining a log of such execution. The
machine can be infected by malware hence there is a good chance of capturing
the malicious behavior in the execution logs, thus allowing researchers to
study such behavior in detail. Although the subsequent analysis of log
information is extensively covered in industrial cybersecurity backends, to our
knowledge there has been only limited effort invested in academia to advance
such log analysis capabilities using cutting edge techniques. We make this
sample dataset available to support designing new machine learning methods for
malware detection, especially for automatic detection of generic malicious
behavior. The dataset has been collected in cooperation between Avast Software
and Czech Technical University - AI Center (AIC). | {
"abstract": "There is a limited amount of publicly available data to support research in\nmalware analysis technology. Particularly, there are virtually no publicly\navailable datasets generated from rich sandboxes such as Cuckoo/CAPE. The\nbenefit of using dynamic sandboxes is the realistic simulation of file\nexecution in the target machine and obtaining a log of such execution. The\nmachine can be infected by malware hence there is a good chance of capturing\nthe malicious behavior in the execution logs, thus allowing researchers to\nstudy such behavior in detail. Although the subsequent analysis of log\ninformation is extensively covered in industrial cybersecurity backends, to our\nknowledge there has been only limited effort invested in academia to advance\nsuch log analysis capabilities using cutting edge techniques. We make this\nsample dataset available to support designing new machine learning methods for\nmalware detection, especially for automatic detection of generic malicious\nbehavior. The dataset has been collected in cooperation between Avast Software\nand Czech Technical University - AI Center (AIC).",
"title": "Avast-CTU Public CAPE Dataset",
"url": "http://arxiv.org/abs/2209.03188v1"
} | null | null | new_dataset | admin | null | false | null | 96829754-3cc8-440a-b232-1ca48c2a3c34 | null | Validated | {
"text_length": 1172
} | 0new_dataset
|
TITLE: RDD2022: A multi-national image dataset for automatic Road Damage Detection
ABSTRACT: The data article describes the Road Damage Dataset, RDD2022, which comprises
47,420 road images from six countries, Japan, India, the Czech Republic,
Norway, the United States, and China. The images have been annotated with more
than 55,000 instances of road damage. Four types of road damage, namely
longitudinal cracks, transverse cracks, alligator cracks, and potholes, are
captured in the dataset. The annotated dataset is envisioned for developing
deep learning-based methods to detect and classify road damage automatically.
The dataset has been released as a part of the Crowd sensing-based Road Damage
Detection Challenge (CRDDC2022). The challenge CRDDC2022 invites researchers
from across the globe to propose solutions for automatic road damage detection
in multiple countries. The municipalities and road agencies may utilize the
RDD2022 dataset, and the models trained using RDD2022 for low-cost automatic
monitoring of road conditions. Further, computer vision and machine learning
researchers may use the dataset to benchmark the performance of different
algorithms for other image-based applications of the same type (classification,
object detection, etc.). | {
"abstract": "The data article describes the Road Damage Dataset, RDD2022, which comprises\n47,420 road images from six countries, Japan, India, the Czech Republic,\nNorway, the United States, and China. The images have been annotated with more\nthan 55,000 instances of road damage. Four types of road damage, namely\nlongitudinal cracks, transverse cracks, alligator cracks, and potholes, are\ncaptured in the dataset. The annotated dataset is envisioned for developing\ndeep learning-based methods to detect and classify road damage automatically.\nThe dataset has been released as a part of the Crowd sensing-based Road Damage\nDetection Challenge (CRDDC2022). The challenge CRDDC2022 invites researchers\nfrom across the globe to propose solutions for automatic road damage detection\nin multiple countries. The municipalities and road agencies may utilize the\nRDD2022 dataset, and the models trained using RDD2022 for low-cost automatic\nmonitoring of road conditions. Further, computer vision and machine learning\nresearchers may use the dataset to benchmark the performance of different\nalgorithms for other image-based applications of the same type (classification,\nobject detection, etc.).",
"title": "RDD2022: A multi-national image dataset for automatic Road Damage Detection",
"url": "http://arxiv.org/abs/2209.08538v1"
} | null | null | new_dataset | admin | null | false | null | c0bcb794-f281-44ee-819a-6bb0253283a0 | null | Validated | {
"text_length": 1284
} | 0new_dataset
|
TITLE: AVOIDDS: Aircraft Vision-based Intruder Detection Dataset and Simulator
ABSTRACT: Designing robust machine learning systems remains an open problem, and there
is a need for benchmark problems that cover both environmental changes and
evaluation on a downstream task. In this work, we introduce AVOIDDS, a
realistic object detection benchmark for the vision-based aircraft
detect-and-avoid problem. We provide a labeled dataset consisting of 72,000
photorealistic images of intruder aircraft with various lighting conditions,
weather conditions, relative geometries, and geographic locations. We also
provide an interface that evaluates trained models on slices of this dataset to
identify changes in performance with respect to changing environmental
conditions. Finally, we implement a fully-integrated, closed-loop simulator of
the vision-based detect-and-avoid problem to evaluate trained models with
respect to the downstream collision avoidance task. This benchmark will enable
further research in the design of robust machine learning systems for use in
safety-critical applications. The AVOIDDS dataset and code are publicly
available at
$\href{https://purl.stanford.edu/hj293cv5980}{purl.stanford.edu/hj293cv5980}$
and
$\href{https://github.com/sisl/VisionBasedAircraftDAA}{github.com/sisl/VisionBasedAircraftDAA}$,
respectively. | {
"abstract": "Designing robust machine learning systems remains an open problem, and there\nis a need for benchmark problems that cover both environmental changes and\nevaluation on a downstream task. In this work, we introduce AVOIDDS, a\nrealistic object detection benchmark for the vision-based aircraft\ndetect-and-avoid problem. We provide a labeled dataset consisting of 72,000\nphotorealistic images of intruder aircraft with various lighting conditions,\nweather conditions, relative geometries, and geographic locations. We also\nprovide an interface that evaluates trained models on slices of this dataset to\nidentify changes in performance with respect to changing environmental\nconditions. Finally, we implement a fully-integrated, closed-loop simulator of\nthe vision-based detect-and-avoid problem to evaluate trained models with\nrespect to the downstream collision avoidance task. This benchmark will enable\nfurther research in the design of robust machine learning systems for use in\nsafety-critical applications. The AVOIDDS dataset and code are publicly\navailable at\n$\\href{https://purl.stanford.edu/hj293cv5980}{purl.stanford.edu/hj293cv5980}$\nand\n$\\href{https://github.com/sisl/VisionBasedAircraftDAA}{github.com/sisl/VisionBasedAircraftDAA}$,\nrespectively.",
"title": "AVOIDDS: Aircraft Vision-based Intruder Detection Dataset and Simulator",
"url": "http://arxiv.org/abs/2306.11203v1"
} | null | null | new_dataset | admin | null | false | null | 1d5acbd8-5051-49fe-b87f-c90562bbe35f | null | Validated | {
"text_length": 1361
} | 0new_dataset
|
TITLE: COMPASS: A Formal Framework and Aggregate Dataset for Generalized Surgical Procedure Modeling
ABSTRACT: Purpose: We propose a formal framework for the modeling and segmentation of
minimally-invasive surgical tasks using a unified set of motion primitives
(MPs) to enable more objective labeling and the aggregation of different
datasets.
Methods: We model dry-lab surgical tasks as finite state machines,
representing how the execution of MPs as the basic surgical actions results in
the change of surgical context, which characterizes the physical interactions
among tools and objects in the surgical environment. We develop methods for
labeling surgical context based on video data and for automatic translation of
context to MP labels. We then use our framework to create the COntext and
Motion Primitive Aggregate Surgical Set (COMPASS), including six dry-lab
surgical tasks from three publicly-available datasets (JIGSAWS, DESK, and
ROSMA), with kinematic and video data and context and MP labels.
Results: Our context labeling method achieves near-perfect agreement between
consensus labels from crowd-sourcing and expert surgeons. Segmentation of tasks
to MPs results in the creation of the COMPASS dataset that nearly triples the
amount of data for modeling and analysis and enables the generation of separate
transcripts for the left and right tools.
Conclusion: The proposed framework results in high quality labeling of
surgical data based on context and fine-grained MPs. Modeling surgical tasks
with MPs enables the aggregation of different datasets and the separate
analysis of left and right hands for bimanual coordination assessment. Our
formal framework and aggregate dataset can support the development of
explainable and multi-granularity models for improved surgical process
analysis, skill assessment, error detection, and autonomy. | {
"abstract": "Purpose: We propose a formal framework for the modeling and segmentation of\nminimally-invasive surgical tasks using a unified set of motion primitives\n(MPs) to enable more objective labeling and the aggregation of different\ndatasets.\n Methods: We model dry-lab surgical tasks as finite state machines,\nrepresenting how the execution of MPs as the basic surgical actions results in\nthe change of surgical context, which characterizes the physical interactions\namong tools and objects in the surgical environment. We develop methods for\nlabeling surgical context based on video data and for automatic translation of\ncontext to MP labels. We then use our framework to create the COntext and\nMotion Primitive Aggregate Surgical Set (COMPASS), including six dry-lab\nsurgical tasks from three publicly-available datasets (JIGSAWS, DESK, and\nROSMA), with kinematic and video data and context and MP labels.\n Results: Our context labeling method achieves near-perfect agreement between\nconsensus labels from crowd-sourcing and expert surgeons. Segmentation of tasks\nto MPs results in the creation of the COMPASS dataset that nearly triples the\namount of data for modeling and analysis and enables the generation of separate\ntranscripts for the left and right tools.\n Conclusion: The proposed framework results in high quality labeling of\nsurgical data based on context and fine-grained MPs. Modeling surgical tasks\nwith MPs enables the aggregation of different datasets and the separate\nanalysis of left and right hands for bimanual coordination assessment. Our\nformal framework and aggregate dataset can support the development of\nexplainable and multi-granularity models for improved surgical process\nanalysis, skill assessment, error detection, and autonomy.",
"title": "COMPASS: A Formal Framework and Aggregate Dataset for Generalized Surgical Procedure Modeling",
"url": "http://arxiv.org/abs/2209.06424v5"
} | null | null | new_dataset | admin | null | false | null | 7ddec2dd-89f4-475d-9328-c02c883bac86 | null | Validated | {
"text_length": 1884
} | 0new_dataset
|
TITLE: DynaBench: A benchmark dataset for learning dynamical systems from low-resolution data
ABSTRACT: Previous work on learning physical systems from data has focused on
high-resolution grid-structured measurements. However, real-world knowledge of
such systems (e.g. weather data) relies on sparsely scattered measuring
stations. In this paper, we introduce a novel simulated benchmark dataset,
DynaBench, for learning dynamical systems directly from sparsely scattered data
without prior knowledge of the equations. The dataset focuses on predicting the
evolution of a dynamical system from low-resolution, unstructured measurements.
We simulate six different partial differential equations covering a variety of
physical systems commonly used in the literature and evaluate several machine
learning models, including traditional graph neural networks and point cloud
processing models, with the task of predicting the evolution of the system. The
proposed benchmark dataset is expected to advance the state of art as an
out-of-the-box easy-to-use tool for evaluating models in a setting where only
unstructured low-resolution observations are available. The benchmark is
available at https://anonymous.4open.science/r/code-2022-dynabench/. | {
"abstract": "Previous work on learning physical systems from data has focused on\nhigh-resolution grid-structured measurements. However, real-world knowledge of\nsuch systems (e.g. weather data) relies on sparsely scattered measuring\nstations. In this paper, we introduce a novel simulated benchmark dataset,\nDynaBench, for learning dynamical systems directly from sparsely scattered data\nwithout prior knowledge of the equations. The dataset focuses on predicting the\nevolution of a dynamical system from low-resolution, unstructured measurements.\nWe simulate six different partial differential equations covering a variety of\nphysical systems commonly used in the literature and evaluate several machine\nlearning models, including traditional graph neural networks and point cloud\nprocessing models, with the task of predicting the evolution of the system. The\nproposed benchmark dataset is expected to advance the state of art as an\nout-of-the-box easy-to-use tool for evaluating models in a setting where only\nunstructured low-resolution observations are available. The benchmark is\navailable at https://anonymous.4open.science/r/code-2022-dynabench/.",
"title": "DynaBench: A benchmark dataset for learning dynamical systems from low-resolution data",
"url": "http://arxiv.org/abs/2306.05805v2"
} | null | null | new_dataset | admin | null | false | null | ea320f3e-d049-455e-adcc-bfbdd1fa8a2b | null | Validated | {
"text_length": 1261
} | 0new_dataset
|
TITLE: Massive MIMO Channel Prediction Via Meta-Learning and Deep Denoising: Is a Small Dataset Enough?
ABSTRACT: Accurate channel knowledge is critical in massive multiple-input
multiple-output (MIMO), which motivates the use of channel prediction. Machine
learning techniques for channel prediction hold much promise, but current
schemes are limited in their ability to adapt to changes in the environment
because they require large training overheads. To accurately predict wireless
channels for new environments with reduced training overhead, we propose a fast
adaptive channel prediction technique based on a meta-learning algorithm for
massive MIMO communications. We exploit the model-agnostic meta-learning (MAML)
algorithm to achieve quick adaptation with a small amount of labeled data.
Also, to improve the prediction accuracy, we adopt the denoising process for
the training data by using deep image prior (DIP). Numerical results show that
the proposed MAML-based channel predictor can improve the prediction accuracy
with only a few fine-tuning samples. The DIP-based denoising process gives an
additional gain in channel prediction, especially in low signal-to-noise ratio
regimes. | {
"abstract": "Accurate channel knowledge is critical in massive multiple-input\nmultiple-output (MIMO), which motivates the use of channel prediction. Machine\nlearning techniques for channel prediction hold much promise, but current\nschemes are limited in their ability to adapt to changes in the environment\nbecause they require large training overheads. To accurately predict wireless\nchannels for new environments with reduced training overhead, we propose a fast\nadaptive channel prediction technique based on a meta-learning algorithm for\nmassive MIMO communications. We exploit the model-agnostic meta-learning (MAML)\nalgorithm to achieve quick adaptation with a small amount of labeled data.\nAlso, to improve the prediction accuracy, we adopt the denoising process for\nthe training data by using deep image prior (DIP). Numerical results show that\nthe proposed MAML-based channel predictor can improve the prediction accuracy\nwith only a few fine-tuning samples. The DIP-based denoising process gives an\nadditional gain in channel prediction, especially in low signal-to-noise ratio\nregimes.",
"title": "Massive MIMO Channel Prediction Via Meta-Learning and Deep Denoising: Is a Small Dataset Enough?",
"url": "http://arxiv.org/abs/2210.08770v1"
} | null | null | no_new_dataset | admin | null | false | null | 60b9cb2d-7bc6-46ad-bc59-6a260bc9a069 | null | Validated | {
"text_length": 1214
} | 1no_new_dataset
|
TITLE: TTSWING: a Dataset for Table Tennis Swing Analysis
ABSTRACT: We introduce TTSWING, a novel dataset designed for table tennis swing
analysis. This dataset comprises comprehensive swing information obtained
through 9-axis sensors integrated into custom-made racket grips, accompanied by
anonymized demographic data of the players. We detail the data collection and
annotation procedures. Furthermore, we conduct pilot studies utilizing diverse
machine learning models for swing analysis. TTSWING holds tremendous potential
to facilitate innovative research in table tennis analysis and is a valuable
resource for the scientific community. We release the dataset and experimental
codes at https://github.com/DEPhantom/TTSWING. | {
"abstract": "We introduce TTSWING, a novel dataset designed for table tennis swing\nanalysis. This dataset comprises comprehensive swing information obtained\nthrough 9-axis sensors integrated into custom-made racket grips, accompanied by\nanonymized demographic data of the players. We detail the data collection and\nannotation procedures. Furthermore, we conduct pilot studies utilizing diverse\nmachine learning models for swing analysis. TTSWING holds tremendous potential\nto facilitate innovative research in table tennis analysis and is a valuable\nresource for the scientific community. We release the dataset and experimental\ncodes at https://github.com/DEPhantom/TTSWING.",
"title": "TTSWING: a Dataset for Table Tennis Swing Analysis",
"url": "http://arxiv.org/abs/2306.17550v1"
} | null | null | new_dataset | admin | null | false | null | 1279b7a2-2a6a-437b-ad7e-bc5a87c03f53 | null | Validated | {
"text_length": 747
} | 0new_dataset
|
TITLE: Evaluating and Crafting Datasets Effective for Deep Learning With Data Maps
ABSTRACT: Rapid development in deep learning model construction has prompted an
increased need for appropriate training data. The popularity of large datasets
- sometimes known as "big data" - has diverted attention from assessing their
quality. Training on large datasets often requires excessive system resources
and an infeasible amount of time. Furthermore, the supervised machine learning
process has yet to be fully automated: for supervised learning, large datasets
require more time for manually labeling samples. We propose a method of
curating smaller datasets with comparable out-of-distribution model accuracy
after an initial training session using an appropriate distribution of samples
classified by how difficult it is for a model to learn from them. | {
"abstract": "Rapid development in deep learning model construction has prompted an\nincreased need for appropriate training data. The popularity of large datasets\n- sometimes known as \"big data\" - has diverted attention from assessing their\nquality. Training on large datasets often requires excessive system resources\nand an infeasible amount of time. Furthermore, the supervised machine learning\nprocess has yet to be fully automated: for supervised learning, large datasets\nrequire more time for manually labeling samples. We propose a method of\ncurating smaller datasets with comparable out-of-distribution model accuracy\nafter an initial training session using an appropriate distribution of samples\nclassified by how difficult it is for a model to learn from them.",
"title": "Evaluating and Crafting Datasets Effective for Deep Learning With Data Maps",
"url": "http://arxiv.org/abs/2208.10033v2"
} | null | null | no_new_dataset | admin | null | false | null | 35163b28-914d-472d-b705-15c31d09c376 | null | Validated | {
"text_length": 866
} | 1no_new_dataset
|
TITLE: Dataset Inference: Ownership Resolution in Machine Learning
ABSTRACT: With increasingly more data and computation involved in their training,
machine learning models constitute valuable intellectual property. This has
spurred interest in model stealing, which is made more practical by advances in
learning with partial, little, or no supervision. Existing defenses focus on
inserting unique watermarks in a model's decision surface, but this is
insufficient: the watermarks are not sampled from the training distribution and
thus are not always preserved during model stealing. In this paper, we make the
key observation that knowledge contained in the stolen model's training set is
what is common to all stolen copies. The adversary's goal, irrespective of the
attack employed, is always to extract this knowledge or its by-products. This
gives the original model's owner a strong advantage over the adversary: model
owners have access to the original training data. We thus introduce $dataset$
$inference$, the process of identifying whether a suspected model copy has
private knowledge from the original model's dataset, as a defense against model
stealing. We develop an approach for dataset inference that combines
statistical testing with the ability to estimate the distance of multiple data
points to the decision boundary. Our experiments on CIFAR10, SVHN, CIFAR100 and
ImageNet show that model owners can claim with confidence greater than 99% that
their model (or dataset as a matter of fact) was stolen, despite only exposing
50 of the stolen model's training points. Dataset inference defends against
state-of-the-art attacks even when the adversary is adaptive. Unlike prior
work, it does not require retraining or overfitting the defended model. | {
"abstract": "With increasingly more data and computation involved in their training,\nmachine learning models constitute valuable intellectual property. This has\nspurred interest in model stealing, which is made more practical by advances in\nlearning with partial, little, or no supervision. Existing defenses focus on\ninserting unique watermarks in a model's decision surface, but this is\ninsufficient: the watermarks are not sampled from the training distribution and\nthus are not always preserved during model stealing. In this paper, we make the\nkey observation that knowledge contained in the stolen model's training set is\nwhat is common to all stolen copies. The adversary's goal, irrespective of the\nattack employed, is always to extract this knowledge or its by-products. This\ngives the original model's owner a strong advantage over the adversary: model\nowners have access to the original training data. We thus introduce $dataset$\n$inference$, the process of identifying whether a suspected model copy has\nprivate knowledge from the original model's dataset, as a defense against model\nstealing. We develop an approach for dataset inference that combines\nstatistical testing with the ability to estimate the distance of multiple data\npoints to the decision boundary. Our experiments on CIFAR10, SVHN, CIFAR100 and\nImageNet show that model owners can claim with confidence greater than 99% that\ntheir model (or dataset as a matter of fact) was stolen, despite only exposing\n50 of the stolen model's training points. Dataset inference defends against\nstate-of-the-art attacks even when the adversary is adaptive. Unlike prior\nwork, it does not require retraining or overfitting the defended model.",
"title": "Dataset Inference: Ownership Resolution in Machine Learning",
"url": "http://arxiv.org/abs/2104.10706v1"
} | null | null | no_new_dataset | admin | null | false | null | 34e24349-1528-48dd-a7cc-b25e36470f4d | null | Validated | {
"text_length": 1786
} | 1no_new_dataset
|
TITLE: A Benchmark dataset for predictive maintenance
ABSTRACT: The paper describes the MetroPT data set, an outcome of a eXplainable
Predictive Maintenance (XPM) project with an urban metro public transportation
service in Porto, Portugal. The data was collected in 2022 that aimed to
evaluate machine learning methods for online anomaly detection and failure
prediction. By capturing several analogic sensor signals (pressure,
temperature, current consumption), digital signals (control signals, discrete
signals), and GPS information (latitude, longitude, and speed), we provide a
dataset that can be easily used to evaluate online machine learning methods.
This dataset contains some interesting characteristics and can be a good
benchmark for predictive maintenance models. | {
"abstract": "The paper describes the MetroPT data set, an outcome of a eXplainable\nPredictive Maintenance (XPM) project with an urban metro public transportation\nservice in Porto, Portugal. The data was collected in 2022 that aimed to\nevaluate machine learning methods for online anomaly detection and failure\nprediction. By capturing several analogic sensor signals (pressure,\ntemperature, current consumption), digital signals (control signals, discrete\nsignals), and GPS information (latitude, longitude, and speed), we provide a\ndataset that can be easily used to evaluate online machine learning methods.\nThis dataset contains some interesting characteristics and can be a good\nbenchmark for predictive maintenance models.",
"title": "A Benchmark dataset for predictive maintenance",
"url": "http://arxiv.org/abs/2207.05466v3"
} | null | null | new_dataset | admin | null | false | null | 35d2af1e-7208-481b-b3eb-491c6f840920 | null | Validated | {
"text_length": 795
} | 0new_dataset
|
TITLE: Data Feedback Loops: Model-driven Amplification of Dataset Biases
ABSTRACT: Datasets scraped from the internet have been critical to the successes of
large-scale machine learning. Yet, this very success puts the utility of future
internet-derived datasets at potential risk, as model outputs begin to replace
human annotations as a source of supervision.
In this work, we first formalize a system where interactions with one model
are recorded as history and scraped as training data in the future. We then
analyze its stability over time by tracking changes to a test-time bias
statistic (e.g. gender bias of model predictions). We find that the degree of
bias amplification is closely linked to whether the model's outputs behave like
samples from the training distribution, a behavior which we characterize and
define as consistent calibration. Experiments in three conditional prediction
scenarios - image classification, visual role-labeling, and language generation
- demonstrate that models that exhibit a sampling-like behavior are more
calibrated and thus more stable. Based on this insight, we propose an
intervention to help calibrate and stabilize unstable feedback systems.
Code is available at https://github.com/rtaori/data_feedback. | {
"abstract": "Datasets scraped from the internet have been critical to the successes of\nlarge-scale machine learning. Yet, this very success puts the utility of future\ninternet-derived datasets at potential risk, as model outputs begin to replace\nhuman annotations as a source of supervision.\n In this work, we first formalize a system where interactions with one model\nare recorded as history and scraped as training data in the future. We then\nanalyze its stability over time by tracking changes to a test-time bias\nstatistic (e.g. gender bias of model predictions). We find that the degree of\nbias amplification is closely linked to whether the model's outputs behave like\nsamples from the training distribution, a behavior which we characterize and\ndefine as consistent calibration. Experiments in three conditional prediction\nscenarios - image classification, visual role-labeling, and language generation\n- demonstrate that models that exhibit a sampling-like behavior are more\ncalibrated and thus more stable. Based on this insight, we propose an\nintervention to help calibrate and stabilize unstable feedback systems.\n Code is available at https://github.com/rtaori/data_feedback.",
"title": "Data Feedback Loops: Model-driven Amplification of Dataset Biases",
"url": "http://arxiv.org/abs/2209.03942v1"
} | null | null | no_new_dataset | admin | null | false | null | e56879a8-ab90-4429-ad5e-20889e984deb | null | Validated | {
"text_length": 1276
} | 1no_new_dataset
|
TITLE: ClimART: A Benchmark Dataset for Emulating Atmospheric Radiative Transfer in Weather and Climate Models
ABSTRACT: Numerical simulations of Earth's weather and climate require substantial
amounts of computation. This has led to a growing interest in replacing
subroutines that explicitly compute physical processes with approximate machine
learning (ML) methods that are fast at inference time. Within weather and
climate models, atmospheric radiative transfer (RT) calculations are especially
expensive. This has made them a popular target for neural network-based
emulators. However, prior work is hard to compare due to the lack of a
comprehensive dataset and standardized best practices for ML benchmarking. To
fill this gap, we build a large dataset, ClimART, with more than \emph{10
million samples from present, pre-industrial, and future climate conditions},
based on the Canadian Earth System Model. ClimART poses several methodological
challenges for the ML community, such as multiple out-of-distribution test
sets, underlying domain physics, and a trade-off between accuracy and inference
speed. We also present several novel baselines that indicate shortcomings of
datasets and network architectures used in prior work. Download instructions,
baselines, and code are available at: https://github.com/RolnickLab/climart | {
"abstract": "Numerical simulations of Earth's weather and climate require substantial\namounts of computation. This has led to a growing interest in replacing\nsubroutines that explicitly compute physical processes with approximate machine\nlearning (ML) methods that are fast at inference time. Within weather and\nclimate models, atmospheric radiative transfer (RT) calculations are especially\nexpensive. This has made them a popular target for neural network-based\nemulators. However, prior work is hard to compare due to the lack of a\ncomprehensive dataset and standardized best practices for ML benchmarking. To\nfill this gap, we build a large dataset, ClimART, with more than \\emph{10\nmillion samples from present, pre-industrial, and future climate conditions},\nbased on the Canadian Earth System Model. ClimART poses several methodological\nchallenges for the ML community, such as multiple out-of-distribution test\nsets, underlying domain physics, and a trade-off between accuracy and inference\nspeed. We also present several novel baselines that indicate shortcomings of\ndatasets and network architectures used in prior work. Download instructions,\nbaselines, and code are available at: https://github.com/RolnickLab/climart",
"title": "ClimART: A Benchmark Dataset for Emulating Atmospheric Radiative Transfer in Weather and Climate Models",
"url": "http://arxiv.org/abs/2111.14671v1"
} | null | null | new_dataset | admin | null | false | null | 279620ec-4c31-4ecb-9f13-d91a3b39ebe6 | null | Validated | {
"text_length": 1354
} | 0new_dataset
|
TITLE: Critical Evaluation of LOCO dataset with Machine Learning
ABSTRACT: Purpose: Object detection is rapidly evolving through machine learning
technology in automation systems. Well prepared data is necessary to train the
algorithms. Accordingly, the objective of this paper is to describe a
re-evaluation of the so-called Logistics Objects in Context (LOCO) dataset,
which is the first dataset for object detection in the field of intralogistics.
Methodology: We use an experimental research approach with three steps to
evaluate the LOCO dataset. Firstly, the images on GitHub were analyzed to
understand the dataset better. Secondly, Google Drive Cloud was used for
training purposes to revisit the algorithmic implementation and training.
Lastly, the LOCO dataset was examined, if it is possible to achieve the same
training results in comparison to the original publications.
Findings: The mean average precision, a common benchmark in object detection,
achieved in our study was 64.54%, and shows a significant increase from the
initial study of the LOCO authors, achieving 41%. However, improvement
potential is seen specifically within object types of forklifts and pallet
truck.
Originality: This paper presents the first critical replication study of the
LOCO dataset for object detection in intralogistics. It shows that the training
with better hyperparameters based on LOCO can even achieve a higher accuracy
than presented in the original publication. However, there is also further room
for improving the LOCO dataset. | {
"abstract": "Purpose: Object detection is rapidly evolving through machine learning\ntechnology in automation systems. Well prepared data is necessary to train the\nalgorithms. Accordingly, the objective of this paper is to describe a\nre-evaluation of the so-called Logistics Objects in Context (LOCO) dataset,\nwhich is the first dataset for object detection in the field of intralogistics.\n Methodology: We use an experimental research approach with three steps to\nevaluate the LOCO dataset. Firstly, the images on GitHub were analyzed to\nunderstand the dataset better. Secondly, Google Drive Cloud was used for\ntraining purposes to revisit the algorithmic implementation and training.\nLastly, the LOCO dataset was examined, if it is possible to achieve the same\ntraining results in comparison to the original publications.\n Findings: The mean average precision, a common benchmark in object detection,\nachieved in our study was 64.54%, and shows a significant increase from the\ninitial study of the LOCO authors, achieving 41%. However, improvement\npotential is seen specifically within object types of forklifts and pallet\ntruck.\n Originality: This paper presents the first critical replication study of the\nLOCO dataset for object detection in intralogistics. It shows that the training\nwith better hyperparameters based on LOCO can even achieve a higher accuracy\nthan presented in the original publication. However, there is also further room\nfor improving the LOCO dataset.",
"title": "Critical Evaluation of LOCO dataset with Machine Learning",
"url": "http://arxiv.org/abs/2209.13499v1"
} | null | null | no_new_dataset | admin | null | false | null | 7892fa27-e42b-4d96-936e-9fd751f74bbe | null | Validated | {
"text_length": 1559
} | 1no_new_dataset
|
TITLE: Architectural Optimization and Feature Learning for High-Dimensional Time Series Datasets
ABSTRACT: As our ability to sense increases, we are experiencing a transition from
data-poor problems, in which the central issue is a lack of relevant data, to
data-rich problems, in which the central issue is to identify a few relevant
features in a sea of observations. Motivated by applications in
gravitational-wave astrophysics, we study the problem of predicting the
presence of transient noise artifacts in a gravitational wave detector from a
rich collection of measurements from the detector and its environment. We argue
that feature learning--in which relevant features are optimized from data--is
critical to achieving high accuracy. We introduce models that reduce the error
rate by over 60% compared to the previous state of the art, which used fixed,
hand-crafted features. Feature learning is useful not only because it improves
performance on prediction tasks; the results provide valuable information about
patterns associated with phenomena of interest that would otherwise be
undiscoverable. In our application, features found to be associated with
transient noise provide diagnostic information about its origin and suggest
mitigation strategies. Learning in high-dimensional settings is challenging.
Through experiments with a variety of architectures, we identify two key
factors in successful models: sparsity, for selecting relevant variables within
the high-dimensional observations; and depth, which confers flexibility for
handling complex interactions and robustness with respect to temporal
variations. We illustrate their significance through systematic experiments on
real detector data. Our results provide experimental corroboration of common
assumptions in the machine-learning community and have direct applicability to
improving our ability to sense gravitational waves, as well as to many other
problem settings with similarly high-dimensional, noisy, or partly irrelevant
data. | {
"abstract": "As our ability to sense increases, we are experiencing a transition from\ndata-poor problems, in which the central issue is a lack of relevant data, to\ndata-rich problems, in which the central issue is to identify a few relevant\nfeatures in a sea of observations. Motivated by applications in\ngravitational-wave astrophysics, we study the problem of predicting the\npresence of transient noise artifacts in a gravitational wave detector from a\nrich collection of measurements from the detector and its environment. We argue\nthat feature learning--in which relevant features are optimized from data--is\ncritical to achieving high accuracy. We introduce models that reduce the error\nrate by over 60% compared to the previous state of the art, which used fixed,\nhand-crafted features. Feature learning is useful not only because it improves\nperformance on prediction tasks; the results provide valuable information about\npatterns associated with phenomena of interest that would otherwise be\nundiscoverable. In our application, features found to be associated with\ntransient noise provide diagnostic information about its origin and suggest\nmitigation strategies. Learning in high-dimensional settings is challenging.\nThrough experiments with a variety of architectures, we identify two key\nfactors in successful models: sparsity, for selecting relevant variables within\nthe high-dimensional observations; and depth, which confers flexibility for\nhandling complex interactions and robustness with respect to temporal\nvariations. We illustrate their significance through systematic experiments on\nreal detector data. Our results provide experimental corroboration of common\nassumptions in the machine-learning community and have direct applicability to\nimproving our ability to sense gravitational waves, as well as to many other\nproblem settings with similarly high-dimensional, noisy, or partly irrelevant\ndata.",
"title": "Architectural Optimization and Feature Learning for High-Dimensional Time Series Datasets",
"url": "http://arxiv.org/abs/2202.13486v2"
} | null | null | no_new_dataset | admin | null | false | null | 354aa25b-0f94-4d9e-b412-a830e2809237 | null | Validated | {
"text_length": 2031
} | 1no_new_dataset
|
TITLE: Potrika: Raw and Balanced Newspaper Datasets in the Bangla Language with Eight Topics and Five Attributes
ABSTRACT: Knowledge is central to human and scientific developments. Natural Language
Processing (NLP) allows automated analysis and creation of knowledge. Data is a
crucial NLP and machine learning ingredient. The scarcity of open datasets is a
well-known problem in machine and deep learning research. This is very much the
case for textual NLP datasets in English and other major world languages. For
the Bangla language, the situation is even more challenging and the number of
large datasets for NLP research is practically nil. We hereby present Potrika,
a large single-label Bangla news article textual dataset curated for NLP
research from six popular online news portals in Bangladesh (Jugantor,
Jaijaidin, Ittefaq, Kaler Kontho, Inqilab, and Somoyer Alo) for the period
2014-2020. The articles are classified into eight distinct categories
(National, Sports, International, Entertainment, Economy, Education, Politics,
and Science \& Technology) providing five attributes (News Article, Category,
Headline, Publication Date, and Newspaper Source). The raw dataset contains
185.51 million words and 12.57 million sentences contained in 664,880 news
articles. Moreover, using NLP augmentation techniques, we create from the raw
(unbalanced) dataset another (balanced) dataset comprising 320,000 news
articles with 40,000 articles in each of the eight news categories. Potrika
contains both the datasets (raw and balanced) to suit a wide range of NLP
research. By far, to the best of our knowledge, Potrika is the largest and the
most extensive dataset for news classification. | {
"abstract": "Knowledge is central to human and scientific developments. Natural Language\nProcessing (NLP) allows automated analysis and creation of knowledge. Data is a\ncrucial NLP and machine learning ingredient. The scarcity of open datasets is a\nwell-known problem in machine and deep learning research. This is very much the\ncase for textual NLP datasets in English and other major world languages. For\nthe Bangla language, the situation is even more challenging and the number of\nlarge datasets for NLP research is practically nil. We hereby present Potrika,\na large single-label Bangla news article textual dataset curated for NLP\nresearch from six popular online news portals in Bangladesh (Jugantor,\nJaijaidin, Ittefaq, Kaler Kontho, Inqilab, and Somoyer Alo) for the period\n2014-2020. The articles are classified into eight distinct categories\n(National, Sports, International, Entertainment, Economy, Education, Politics,\nand Science \\& Technology) providing five attributes (News Article, Category,\nHeadline, Publication Date, and Newspaper Source). The raw dataset contains\n185.51 million words and 12.57 million sentences contained in 664,880 news\narticles. Moreover, using NLP augmentation techniques, we create from the raw\n(unbalanced) dataset another (balanced) dataset comprising 320,000 news\narticles with 40,000 articles in each of the eight news categories. Potrika\ncontains both the datasets (raw and balanced) to suit a wide range of NLP\nresearch. By far, to the best of our knowledge, Potrika is the largest and the\nmost extensive dataset for news classification.",
"title": "Potrika: Raw and Balanced Newspaper Datasets in the Bangla Language with Eight Topics and Five Attributes",
"url": "http://arxiv.org/abs/2210.09389v1"
} | null | null | new_dataset | admin | null | false | null | 7fe08fc6-104f-4522-b868-4114f14aa33c | null | Validated | {
"text_length": 1714
} | 0new_dataset
|
TITLE: The Open Catalyst 2022 (OC22) Dataset and Challenges for Oxide Electrocatalysts
ABSTRACT: The development of machine learning models for electrocatalysts requires a
broad set of training data to enable their use across a wide variety of
materials. One class of materials that currently lacks sufficient training data
is oxides, which are critical for the development of OER catalysts. To address
this, we developed the OC22 dataset, consisting of 62,331 DFT relaxations
(~9,854,504 single point calculations) across a range of oxide materials,
coverages, and adsorbates. We define generalized total energy tasks that enable
property prediction beyond adsorption energies; we test baseline performance of
several graph neural networks; and we provide pre-defined dataset splits to
establish clear benchmarks for future efforts. In the most general task,
GemNet-OC sees a ~36% improvement in energy predictions when combining the
chemically dissimilar OC20 and OC22 datasets via fine-tuning. Similarly, we
achieved a ~19% improvement in total energy predictions on OC20 and a ~9%
improvement in force predictions in OC22 when using joint training. We
demonstrate the practical utility of a top performing model by capturing
literature adsorption energies and important OER scaling relationships. We
expect OC22 to provide an important benchmark for models seeking to incorporate
intricate long-range electrostatic and magnetic interactions in oxide surfaces.
Dataset and baseline models are open sourced, and a public leaderboard is
available to encourage continued community developments on the total energy
tasks and data. | {
"abstract": "The development of machine learning models for electrocatalysts requires a\nbroad set of training data to enable their use across a wide variety of\nmaterials. One class of materials that currently lacks sufficient training data\nis oxides, which are critical for the development of OER catalysts. To address\nthis, we developed the OC22 dataset, consisting of 62,331 DFT relaxations\n(~9,854,504 single point calculations) across a range of oxide materials,\ncoverages, and adsorbates. We define generalized total energy tasks that enable\nproperty prediction beyond adsorption energies; we test baseline performance of\nseveral graph neural networks; and we provide pre-defined dataset splits to\nestablish clear benchmarks for future efforts. In the most general task,\nGemNet-OC sees a ~36% improvement in energy predictions when combining the\nchemically dissimilar OC20 and OC22 datasets via fine-tuning. Similarly, we\nachieved a ~19% improvement in total energy predictions on OC20 and a ~9%\nimprovement in force predictions in OC22 when using joint training. We\ndemonstrate the practical utility of a top performing model by capturing\nliterature adsorption energies and important OER scaling relationships. We\nexpect OC22 to provide an important benchmark for models seeking to incorporate\nintricate long-range electrostatic and magnetic interactions in oxide surfaces.\nDataset and baseline models are open sourced, and a public leaderboard is\navailable to encourage continued community developments on the total energy\ntasks and data.",
"title": "The Open Catalyst 2022 (OC22) Dataset and Challenges for Oxide Electrocatalysts",
"url": "http://arxiv.org/abs/2206.08917v3"
} | null | null | new_dataset | admin | null | false | null | 165e43bb-eb0b-4bc5-a64e-a4cdf40f12c2 | null | Validated | {
"text_length": 1646
} | 0new_dataset
|
TITLE: BrazilDAM: A Benchmark dataset for Tailings Dam Detection
ABSTRACT: In this work we present BrazilDAM, a novel public dataset based on Sentinel-2
and Landsat-8 satellite images covering all tailings dams cataloged by the
Brazilian National Mining Agency (ANM). The dataset was built using
georeferenced images from 769 dams, recorded between 2016 and 2019. The time
series were processed in order to produce cloud free images. The dams contain
mining waste from different ore categories and have highly varying shapes,
areas and volumes, making BrazilDAM particularly interesting and challenging to
be used in machine learning benchmarks. The original catalog contains, besides
the dam coordinates, information about: the main ore, constructive method, risk
category, and associated potential damage. To evaluate BrazilDAM's predictive
potential we performed classification essays using state-of-the-art deep
Convolutional Neural Network (CNNs). In the experiments, we achieved an average
classification accuracy of 94.11% in tailing dam binary classification task. In
addition, others four setups of experiments were made using the complementary
information from the original catalog, exhaustively exploiting the capacity of
the proposed dataset. | {
"abstract": "In this work we present BrazilDAM, a novel public dataset based on Sentinel-2\nand Landsat-8 satellite images covering all tailings dams cataloged by the\nBrazilian National Mining Agency (ANM). The dataset was built using\ngeoreferenced images from 769 dams, recorded between 2016 and 2019. The time\nseries were processed in order to produce cloud free images. The dams contain\nmining waste from different ore categories and have highly varying shapes,\nareas and volumes, making BrazilDAM particularly interesting and challenging to\nbe used in machine learning benchmarks. The original catalog contains, besides\nthe dam coordinates, information about: the main ore, constructive method, risk\ncategory, and associated potential damage. To evaluate BrazilDAM's predictive\npotential we performed classification essays using state-of-the-art deep\nConvolutional Neural Network (CNNs). In the experiments, we achieved an average\nclassification accuracy of 94.11% in tailing dam binary classification task. In\naddition, others four setups of experiments were made using the complementary\ninformation from the original catalog, exhaustively exploiting the capacity of\nthe proposed dataset.",
"title": "BrazilDAM: A Benchmark dataset for Tailings Dam Detection",
"url": "http://arxiv.org/abs/2003.07948v2"
} | null | null | new_dataset | admin | null | false | null | df34d82e-ae49-42db-b2cc-daf65dee09d6 | null | Validated | {
"text_length": 1271
} | 0new_dataset
|
TITLE: Generative Adversarial Nets: Can we generate a new dataset based on only one training set?
ABSTRACT: A generative adversarial network (GAN) is a class of machine learning
frameworks designed by Goodfellow et al. in 2014. In the GAN framework, the
generative model is pitted against an adversary: a discriminative model that
learns to determine whether a sample is from the model distribution or the data
distribution. GAN generates new samples from the same distribution as the
training set. In this work, we aim to generate a new dataset that has a
different distribution from the training set. In addition, the Jensen-Shannon
divergence between the distributions of the generative and training datasets
can be controlled by some target $\delta \in [0, 1]$. Our work is motivated by
applications in generating new kinds of rice that have similar characteristics
as good rice. | {
"abstract": "A generative adversarial network (GAN) is a class of machine learning\nframeworks designed by Goodfellow et al. in 2014. In the GAN framework, the\ngenerative model is pitted against an adversary: a discriminative model that\nlearns to determine whether a sample is from the model distribution or the data\ndistribution. GAN generates new samples from the same distribution as the\ntraining set. In this work, we aim to generate a new dataset that has a\ndifferent distribution from the training set. In addition, the Jensen-Shannon\ndivergence between the distributions of the generative and training datasets\ncan be controlled by some target $\\delta \\in [0, 1]$. Our work is motivated by\napplications in generating new kinds of rice that have similar characteristics\nas good rice.",
"title": "Generative Adversarial Nets: Can we generate a new dataset based on only one training set?",
"url": "http://arxiv.org/abs/2210.06005v1"
} | null | null | no_new_dataset | admin | null | false | null | 52c20045-03c7-4b3d-a2db-b5f74019896c | null | Default | {
"text_length": 900
} | 1no_new_dataset
|
TITLE: Synthetic Dataset Generation for Privacy-Preserving Machine Learning
ABSTRACT: Machine Learning (ML) has achieved enormous success in solving a variety of
problems in computer vision, speech recognition, object detection, to name a
few. The principal reason for this success is the availability of huge datasets
for training deep neural networks (DNNs). However, datasets can not be publicly
released if they contain sensitive information such as medical or financial
records. In such cases, data privacy becomes a major concern. Encryption
methods offer a possible solution to this issue, however their deployment on ML
applications is non-trivial, as they seriously impact the classification
accuracy and result in substantial computational overhead.Alternatively,
obfuscation techniques can be used, but maintaining a good balance between
visual privacy and accuracy is challenging. In this work, we propose a method
to generate secure synthetic datasets from the original private datasets. In
our method, given a network with Batch Normalization (BN) layers pre-trained on
the original dataset, we first record the layer-wise BN statistics. Next, using
the BN statistics and the pre-trained model, we generate the synthetic dataset
by optimizing random noises such that the synthetic data match the layer-wise
statistical distribution of the original model. We evaluate our method on image
classification dataset (CIFAR10) and show that our synthetic data can be used
for training networks from scratch, producing reasonable classification
performance. | {
"abstract": "Machine Learning (ML) has achieved enormous success in solving a variety of\nproblems in computer vision, speech recognition, object detection, to name a\nfew. The principal reason for this success is the availability of huge datasets\nfor training deep neural networks (DNNs). However, datasets can not be publicly\nreleased if they contain sensitive information such as medical or financial\nrecords. In such cases, data privacy becomes a major concern. Encryption\nmethods offer a possible solution to this issue, however their deployment on ML\napplications is non-trivial, as they seriously impact the classification\naccuracy and result in substantial computational overhead.Alternatively,\nobfuscation techniques can be used, but maintaining a good balance between\nvisual privacy and accuracy is challenging. In this work, we propose a method\nto generate secure synthetic datasets from the original private datasets. In\nour method, given a network with Batch Normalization (BN) layers pre-trained on\nthe original dataset, we first record the layer-wise BN statistics. Next, using\nthe BN statistics and the pre-trained model, we generate the synthetic dataset\nby optimizing random noises such that the synthetic data match the layer-wise\nstatistical distribution of the original model. We evaluate our method on image\nclassification dataset (CIFAR10) and show that our synthetic data can be used\nfor training networks from scratch, producing reasonable classification\nperformance.",
"title": "Synthetic Dataset Generation for Privacy-Preserving Machine Learning",
"url": "http://arxiv.org/abs/2210.03205v5"
} | null | null | no_new_dataset | admin | null | false | null | 7aea2fea-96db-4732-8496-7f4da428882c | null | Validated | {
"text_length": 1580
} | 1no_new_dataset
|
TITLE: An Instance Selection Algorithm for Big Data in High imbalanced datasets based on LSH
ABSTRACT: Training of Machine Learning (ML) models in real contexts often deals with
big data sets and high-class imbalance samples where the class of interest is
unrepresented (minority class). Practical solutions using classical ML models
address the problem of large data sets using parallel/distributed
implementations of training algorithms, approximate model-based solutions, or
applying instance selection (IS) algorithms to eliminate redundant information.
However, the combined problem of big and high imbalanced datasets has been less
addressed. This work proposes three new methods for IS to be able to deal with
large and imbalanced data sets. The proposed methods use Locality Sensitive
Hashing (LSH) as a base clustering technique, and then three different sampling
methods are applied on top of the clusters (or buckets) generated by LSH. The
algorithms were developed in the Apache Spark framework, guaranteeing their
scalability. The experiments carried out in three different datasets suggest
that the proposed IS methods can improve the performance of a base ML model
between 5% and 19% in terms of the geometric mean. | {
"abstract": "Training of Machine Learning (ML) models in real contexts often deals with\nbig data sets and high-class imbalance samples where the class of interest is\nunrepresented (minority class). Practical solutions using classical ML models\naddress the problem of large data sets using parallel/distributed\nimplementations of training algorithms, approximate model-based solutions, or\napplying instance selection (IS) algorithms to eliminate redundant information.\nHowever, the combined problem of big and high imbalanced datasets has been less\naddressed. This work proposes three new methods for IS to be able to deal with\nlarge and imbalanced data sets. The proposed methods use Locality Sensitive\nHashing (LSH) as a base clustering technique, and then three different sampling\nmethods are applied on top of the clusters (or buckets) generated by LSH. The\nalgorithms were developed in the Apache Spark framework, guaranteeing their\nscalability. The experiments carried out in three different datasets suggest\nthat the proposed IS methods can improve the performance of a base ML model\nbetween 5% and 19% in terms of the geometric mean.",
"title": "An Instance Selection Algorithm for Big Data in High imbalanced datasets based on LSH",
"url": "http://arxiv.org/abs/2210.04310v1"
} | null | null | no_new_dataset | admin | null | false | null | 2587e06f-716b-4f13-9596-bfcabfb1e988 | null | Validated | {
"text_length": 1247
} | 1no_new_dataset
|
TITLE: Classification of datasets with imputed missing values: does imputation quality matter?
ABSTRACT: Classifying samples in incomplete datasets is a common aim for machine
learning practitioners, but is non-trivial. Missing data is found in most
real-world datasets and these missing values are typically imputed using
established methods, followed by classification of the now complete, imputed,
samples. The focus of the machine learning researcher is then to optimise the
downstream classification performance. In this study, we highlight that it is
imperative to consider the quality of the imputation. We demonstrate how the
commonly used measures for assessing quality are flawed and propose a new class
of discrepancy scores which focus on how well the method recreates the overall
distribution of the data. To conclude, we highlight the compromised
interpretability of classifier models trained using poorly imputed data. | {
"abstract": "Classifying samples in incomplete datasets is a common aim for machine\nlearning practitioners, but is non-trivial. Missing data is found in most\nreal-world datasets and these missing values are typically imputed using\nestablished methods, followed by classification of the now complete, imputed,\nsamples. The focus of the machine learning researcher is then to optimise the\ndownstream classification performance. In this study, we highlight that it is\nimperative to consider the quality of the imputation. We demonstrate how the\ncommonly used measures for assessing quality are flawed and propose a new class\nof discrepancy scores which focus on how well the method recreates the overall\ndistribution of the data. To conclude, we highlight the compromised\ninterpretability of classifier models trained using poorly imputed data.",
"title": "Classification of datasets with imputed missing values: does imputation quality matter?",
"url": "http://arxiv.org/abs/2206.08478v1"
} | null | null | no_new_dataset | admin | null | false | null | 2647bd44-498b-42f9-9ef8-6cb20b36eed9 | null | Validated | {
"text_length": 950
} | 1no_new_dataset
|
TITLE: Scalable mRMR feature selection to handle high dimensional datasets: Vertical partitioning based Iterative MapReduce framework
ABSTRACT: While building machine learning models, Feature selection (FS) stands out as
an essential preprocessing step used to handle the uncertainty and vagueness in
the data. Recently, the minimum Redundancy and Maximum Relevance (mRMR)
approach has proven to be effective in obtaining the irredundant feature
subset. Owing to the generation of voluminous datasets, it is essential to
design scalable solutions using distributed/parallel paradigms. MapReduce
solutions are proven to be one of the best approaches to designing
fault-tolerant and scalable solutions. This work analyses the existing
MapReduce approaches for mRMR feature selection and identifies the limitations
thereof. In the current study, we proposed VMR_mRMR, an efficient vertical
partitioning-based approach using a memorization approach, thereby overcoming
the extant approaches limitations. The experiment analysis says that VMR_mRMR
significantly outperformed extant approaches and achieved a better
computational gain (C.G). In addition, we also conducted a comparative analysis
with the horizontal partitioning approach HMR_mRMR [1] to assess the strengths
and limitations of the proposed approach. | {
"abstract": "While building machine learning models, Feature selection (FS) stands out as\nan essential preprocessing step used to handle the uncertainty and vagueness in\nthe data. Recently, the minimum Redundancy and Maximum Relevance (mRMR)\napproach has proven to be effective in obtaining the irredundant feature\nsubset. Owing to the generation of voluminous datasets, it is essential to\ndesign scalable solutions using distributed/parallel paradigms. MapReduce\nsolutions are proven to be one of the best approaches to designing\nfault-tolerant and scalable solutions. This work analyses the existing\nMapReduce approaches for mRMR feature selection and identifies the limitations\nthereof. In the current study, we proposed VMR_mRMR, an efficient vertical\npartitioning-based approach using a memorization approach, thereby overcoming\nthe extant approaches limitations. The experiment analysis says that VMR_mRMR\nsignificantly outperformed extant approaches and achieved a better\ncomputational gain (C.G). In addition, we also conducted a comparative analysis\nwith the horizontal partitioning approach HMR_mRMR [1] to assess the strengths\nand limitations of the proposed approach.",
"title": "Scalable mRMR feature selection to handle high dimensional datasets: Vertical partitioning based Iterative MapReduce framework",
"url": "http://arxiv.org/abs/2208.09901v1"
} | null | null | no_new_dataset | admin | null | false | null | 7ce72dae-c96a-4d20-b0ed-9d4e1476739b | null | Validated | {
"text_length": 1327
} | 1no_new_dataset
|
TITLE: Deep Learning based Monocular Depth Prediction: Datasets, Methods and Applications
ABSTRACT: Estimating depth from RGB images can facilitate many computer vision tasks,
such as indoor localization, height estimation, and simultaneous localization
and mapping (SLAM). Recently, monocular depth estimation has obtained great
progress owing to the rapid development of deep learning techniques. They
surpass traditional machine learning-based methods by a large margin in terms
of accuracy and speed. Despite the rapid progress in this topic, there are
lacking of a comprehensive review, which is needed to summarize the current
progress and provide the future directions. In this survey, we first introduce
the datasets for depth estimation, and then give a comprehensive introduction
of the methods from three perspectives: supervised learning-based methods,
unsupervised learning-based methods, and sparse samples guidance-based methods.
In addition, downstream applications that benefit from the progress have also
been illustrated. Finally, we point out the future directions and conclude the
paper. | {
"abstract": "Estimating depth from RGB images can facilitate many computer vision tasks,\nsuch as indoor localization, height estimation, and simultaneous localization\nand mapping (SLAM). Recently, monocular depth estimation has obtained great\nprogress owing to the rapid development of deep learning techniques. They\nsurpass traditional machine learning-based methods by a large margin in terms\nof accuracy and speed. Despite the rapid progress in this topic, there are\nlacking of a comprehensive review, which is needed to summarize the current\nprogress and provide the future directions. In this survey, we first introduce\nthe datasets for depth estimation, and then give a comprehensive introduction\nof the methods from three perspectives: supervised learning-based methods,\nunsupervised learning-based methods, and sparse samples guidance-based methods.\nIn addition, downstream applications that benefit from the progress have also\nbeen illustrated. Finally, we point out the future directions and conclude the\npaper.",
"title": "Deep Learning based Monocular Depth Prediction: Datasets, Methods and Applications",
"url": "http://arxiv.org/abs/2011.04123v1"
} | null | null | no_new_dataset | admin | null | false | null | 27cc65f3-37c2-4ed9-aa19-4a91779f034f | null | Validated | {
"text_length": 1125
} | 1no_new_dataset
|
TITLE: MRCLens: an MRC Dataset Bias Detection Toolkit
ABSTRACT: Many recent neural models have shown remarkable empirical results in Machine
Reading Comprehension, but evidence suggests sometimes the models take
advantage of dataset biases to predict and fail to generalize on out-of-sample
data. While many other approaches have been proposed to address this issue from
the computation perspective such as new architectures or training procedures,
we believe a method that allows researchers to discover biases, and adjust the
data or the models in an earlier stage will be beneficial. Thus, we introduce
MRCLens, a toolkit that detects whether biases exist before users train the
full model. For the convenience of introducing the toolkit, we also provide a
categorization of common biases in MRC. | {
"abstract": "Many recent neural models have shown remarkable empirical results in Machine\nReading Comprehension, but evidence suggests sometimes the models take\nadvantage of dataset biases to predict and fail to generalize on out-of-sample\ndata. While many other approaches have been proposed to address this issue from\nthe computation perspective such as new architectures or training procedures,\nwe believe a method that allows researchers to discover biases, and adjust the\ndata or the models in an earlier stage will be beneficial. Thus, we introduce\nMRCLens, a toolkit that detects whether biases exist before users train the\nfull model. For the convenience of introducing the toolkit, we also provide a\ncategorization of common biases in MRC.",
"title": "MRCLens: an MRC Dataset Bias Detection Toolkit",
"url": "http://arxiv.org/abs/2207.08943v1"
} | null | null | no_new_dataset | admin | null | false | null | 298e2a99-0e5b-49a9-935b-ddb37e83be36 | null | Validated | {
"text_length": 816
} | 1no_new_dataset
|
TITLE: Ex-Ante Assessment of Discrimination in Dataset
ABSTRACT: Data owners face increasing liability for how the use of their data could
harm under-priviliged communities. Stakeholders would like to identify the
characteristics of data that lead to algorithms being biased against any
particular demographic groups, for example, defined by their race, gender, age,
and/or religion. Specifically, we are interested in identifying subsets of the
feature space where the ground truth response function from features to
observed outcomes differs across demographic groups. To this end, we propose
FORESEE, a FORESt of decision trEEs algorithm, which generates a score that
captures how likely an individual's response varies with sensitive attributes.
Empirically, we find that our approach allows us to identify the individuals
who are most likely to be misclassified by several classifiers, including
Random Forest, Logistic Regression, Support Vector Machine, and k-Nearest
Neighbors. The advantage of our approach is that it allows stakeholders to
characterize risky samples that may contribute to discrimination, as well as,
use the FORESEE to estimate the risk of upcoming samples. | {
"abstract": "Data owners face increasing liability for how the use of their data could\nharm under-priviliged communities. Stakeholders would like to identify the\ncharacteristics of data that lead to algorithms being biased against any\nparticular demographic groups, for example, defined by their race, gender, age,\nand/or religion. Specifically, we are interested in identifying subsets of the\nfeature space where the ground truth response function from features to\nobserved outcomes differs across demographic groups. To this end, we propose\nFORESEE, a FORESt of decision trEEs algorithm, which generates a score that\ncaptures how likely an individual's response varies with sensitive attributes.\nEmpirically, we find that our approach allows us to identify the individuals\nwho are most likely to be misclassified by several classifiers, including\nRandom Forest, Logistic Regression, Support Vector Machine, and k-Nearest\nNeighbors. The advantage of our approach is that it allows stakeholders to\ncharacterize risky samples that may contribute to discrimination, as well as,\nuse the FORESEE to estimate the risk of upcoming samples.",
"title": "Ex-Ante Assessment of Discrimination in Dataset",
"url": "http://arxiv.org/abs/2208.07918v2"
} | null | null | no_new_dataset | admin | null | false | null | 371118da-b9ee-480b-9a75-c06b8cb1717e | null | Validated | {
"text_length": 1202
} | 1no_new_dataset
|
TITLE: Gender and Racial Bias in Visual Question Answering Datasets
ABSTRACT: Vision-and-language tasks have increasingly drawn more attention as a means
to evaluate human-like reasoning in machine learning models. A popular task in
the field is visual question answering (VQA), which aims to answer questions
about images. However, VQA models have been shown to exploit language bias by
learning the statistical correlations between questions and answers without
looking into the image content: e.g., questions about the color of a banana are
answered with yellow, even if the banana in the image is green. If societal
bias (e.g., sexism, racism, ableism, etc.) is present in the training data,
this problem may be causing VQA models to learn harmful stereotypes. For this
reason, we investigate gender and racial bias in five VQA datasets. In our
analysis, we find that the distribution of answers is highly different between
questions about women and men, as well as the existence of detrimental
gender-stereotypical samples. Likewise, we identify that specific race-related
attributes are underrepresented, whereas potentially discriminatory samples
appear in the analyzed datasets. Our findings suggest that there are dangers
associated to using VQA datasets without considering and dealing with the
potentially harmful stereotypes. We conclude the paper by proposing solutions
to alleviate the problem before, during, and after the dataset collection
process. | {
"abstract": "Vision-and-language tasks have increasingly drawn more attention as a means\nto evaluate human-like reasoning in machine learning models. A popular task in\nthe field is visual question answering (VQA), which aims to answer questions\nabout images. However, VQA models have been shown to exploit language bias by\nlearning the statistical correlations between questions and answers without\nlooking into the image content: e.g., questions about the color of a banana are\nanswered with yellow, even if the banana in the image is green. If societal\nbias (e.g., sexism, racism, ableism, etc.) is present in the training data,\nthis problem may be causing VQA models to learn harmful stereotypes. For this\nreason, we investigate gender and racial bias in five VQA datasets. In our\nanalysis, we find that the distribution of answers is highly different between\nquestions about women and men, as well as the existence of detrimental\ngender-stereotypical samples. Likewise, we identify that specific race-related\nattributes are underrepresented, whereas potentially discriminatory samples\nappear in the analyzed datasets. Our findings suggest that there are dangers\nassociated to using VQA datasets without considering and dealing with the\npotentially harmful stereotypes. We conclude the paper by proposing solutions\nto alleviate the problem before, during, and after the dataset collection\nprocess.",
"title": "Gender and Racial Bias in Visual Question Answering Datasets",
"url": "http://arxiv.org/abs/2205.08148v3"
} | null | null | no_new_dataset | admin | null | false | null | a2dc7836-2bbd-4b76-b785-1e06276846c0 | null | Validated | {
"text_length": 1482
} | 1no_new_dataset
|
TITLE: Computer Vision based inspection on post-earthquake with UAV synthetic dataset
ABSTRACT: The area affected by the earthquake is vast and often difficult to entirely
cover, and the earthquake itself is a sudden event that causes multiple defects
simultaneously, that cannot be effectively traced using traditional, manual
methods. This article presents an innovative approach to the problem of
detecting damage after sudden events by using an interconnected set of deep
machine learning models organized in a single pipeline and allowing for easy
modification and swapping models seamlessly. Models in the pipeline were
trained with a synthetic dataset and were adapted to be further evaluated and
used with unmanned aerial vehicles (UAVs) in real-world conditions. Thanks to
the methods presented in the article, it is possible to obtain high accuracy in
detecting buildings defects, segmenting constructions into their components and
estimating their technical condition based on a single drone flight. | {
"abstract": "The area affected by the earthquake is vast and often difficult to entirely\ncover, and the earthquake itself is a sudden event that causes multiple defects\nsimultaneously, that cannot be effectively traced using traditional, manual\nmethods. This article presents an innovative approach to the problem of\ndetecting damage after sudden events by using an interconnected set of deep\nmachine learning models organized in a single pipeline and allowing for easy\nmodification and swapping models seamlessly. Models in the pipeline were\ntrained with a synthetic dataset and were adapted to be further evaluated and\nused with unmanned aerial vehicles (UAVs) in real-world conditions. Thanks to\nthe methods presented in the article, it is possible to obtain high accuracy in\ndetecting buildings defects, segmenting constructions into their components and\nestimating their technical condition based on a single drone flight.",
"title": "Computer Vision based inspection on post-earthquake with UAV synthetic dataset",
"url": "http://arxiv.org/abs/2210.05282v1"
} | null | null | no_new_dataset | admin | null | false | null | 436923c6-aa42-45c0-a8a4-7e6e9dbb60e4 | null | Validated | {
"text_length": 1027
} | 1no_new_dataset
|
TITLE: HAPI: A Large-scale Longitudinal Dataset of Commercial ML API Predictions
ABSTRACT: Commercial ML APIs offered by providers such as Google, Amazon and Microsoft
have dramatically simplified ML adoption in many applications. Numerous
companies and academics pay to use ML APIs for tasks such as object detection,
OCR and sentiment analysis. Different ML APIs tackling the same task can have
very heterogeneous performance. Moreover, the ML models underlying the APIs
also evolve over time. As ML APIs rapidly become a valuable marketplace and a
widespread way to consume machine learning, it is critical to systematically
study and compare different APIs with each other and to characterize how APIs
change over time. However, this topic is currently underexplored due to the
lack of data. In this paper, we present HAPI (History of APIs), a longitudinal
dataset of 1,761,417 instances of commercial ML API applications (involving
APIs from Amazon, Google, IBM, Microsoft and other providers) across diverse
tasks including image tagging, speech recognition and text mining from 2020 to
2022. Each instance consists of a query input for an API (e.g., an image or
text) along with the API's output prediction/annotation and confidence scores.
HAPI is the first large-scale dataset of ML API usages and is a unique resource
for studying ML-as-a-service (MLaaS). As examples of the types of analyses that
HAPI enables, we show that ML APIs' performance change substantially over
time--several APIs' accuracies dropped on specific benchmark datasets. Even
when the API's aggregate performance stays steady, its error modes can shift
across different subtypes of data between 2020 and 2022. Such changes can
substantially impact the entire analytics pipelines that use some ML API as a
component. We further use HAPI to study commercial APIs' performance
disparities across demographic subgroups over time. HAPI can stimulate more
research in the growing field of MLaaS. | {
"abstract": "Commercial ML APIs offered by providers such as Google, Amazon and Microsoft\nhave dramatically simplified ML adoption in many applications. Numerous\ncompanies and academics pay to use ML APIs for tasks such as object detection,\nOCR and sentiment analysis. Different ML APIs tackling the same task can have\nvery heterogeneous performance. Moreover, the ML models underlying the APIs\nalso evolve over time. As ML APIs rapidly become a valuable marketplace and a\nwidespread way to consume machine learning, it is critical to systematically\nstudy and compare different APIs with each other and to characterize how APIs\nchange over time. However, this topic is currently underexplored due to the\nlack of data. In this paper, we present HAPI (History of APIs), a longitudinal\ndataset of 1,761,417 instances of commercial ML API applications (involving\nAPIs from Amazon, Google, IBM, Microsoft and other providers) across diverse\ntasks including image tagging, speech recognition and text mining from 2020 to\n2022. Each instance consists of a query input for an API (e.g., an image or\ntext) along with the API's output prediction/annotation and confidence scores.\nHAPI is the first large-scale dataset of ML API usages and is a unique resource\nfor studying ML-as-a-service (MLaaS). As examples of the types of analyses that\nHAPI enables, we show that ML APIs' performance change substantially over\ntime--several APIs' accuracies dropped on specific benchmark datasets. Even\nwhen the API's aggregate performance stays steady, its error modes can shift\nacross different subtypes of data between 2020 and 2022. Such changes can\nsubstantially impact the entire analytics pipelines that use some ML API as a\ncomponent. We further use HAPI to study commercial APIs' performance\ndisparities across demographic subgroups over time. HAPI can stimulate more\nresearch in the growing field of MLaaS.",
"title": "HAPI: A Large-scale Longitudinal Dataset of Commercial ML API Predictions",
"url": "http://arxiv.org/abs/2209.08443v1"
} | null | null | new_dataset | admin | null | false | null | dbb8c2fe-bb43-4e45-9fd7-15a622ece478 | null | Validated | {
"text_length": 1988
} | 0new_dataset
|
TITLE: GLARE: A Dataset for Traffic Sign Detection in Sun Glare
ABSTRACT: Real-time machine learning detection algorithms are often found within
autonomous vehicle technology and depend on quality datasets. It is essential
that these algorithms work correctly in everyday conditions as well as under
strong sun glare. Reports indicate glare is one of the two most prominent
environment-related reasons for crashes. However, existing datasets, such as
LISA and the German Traffic Sign Recognition Benchmark, do not reflect the
existence of sun glare at all. This paper presents the GLARE traffic sign
dataset: a collection of images with U.S based traffic signs under heavy visual
interference by sunlight. GLARE contains 2,157 images of traffic signs with sun
glare, pulled from 33 videos of dashcam footage of roads in the United States.
It provides an essential enrichment to the widely used LISA Traffic Sign
dataset. Our experimental study shows that although several state-of-the-art
baseline methods demonstrate superior performance when trained and tested
against traffic sign datasets without sun glare, they greatly suffer when
tested against GLARE (e.g., ranging from 9% to 21% mean mAP, which is
significantly lower than the performances on LISA dataset). We also notice that
current architectures have better detection accuracy (e.g., on average 42% mean
mAP gain for mainstream algorithms) when trained on images of traffic signs in
sun glare. | {
"abstract": "Real-time machine learning detection algorithms are often found within\nautonomous vehicle technology and depend on quality datasets. It is essential\nthat these algorithms work correctly in everyday conditions as well as under\nstrong sun glare. Reports indicate glare is one of the two most prominent\nenvironment-related reasons for crashes. However, existing datasets, such as\nLISA and the German Traffic Sign Recognition Benchmark, do not reflect the\nexistence of sun glare at all. This paper presents the GLARE traffic sign\ndataset: a collection of images with U.S based traffic signs under heavy visual\ninterference by sunlight. GLARE contains 2,157 images of traffic signs with sun\nglare, pulled from 33 videos of dashcam footage of roads in the United States.\nIt provides an essential enrichment to the widely used LISA Traffic Sign\ndataset. Our experimental study shows that although several state-of-the-art\nbaseline methods demonstrate superior performance when trained and tested\nagainst traffic sign datasets without sun glare, they greatly suffer when\ntested against GLARE (e.g., ranging from 9% to 21% mean mAP, which is\nsignificantly lower than the performances on LISA dataset). We also notice that\ncurrent architectures have better detection accuracy (e.g., on average 42% mean\nmAP gain for mainstream algorithms) when trained on images of traffic signs in\nsun glare.",
"title": "GLARE: A Dataset for Traffic Sign Detection in Sun Glare",
"url": "http://arxiv.org/abs/2209.08716v1"
} | null | null | new_dataset | admin | null | false | null | 8a57c963-93a0-4088-8f7b-93262692f72a | null | Validated | {
"text_length": 1473
} | 0new_dataset
|
TITLE: COVIDx CT-3: A Large-scale, Multinational, Open-Source Benchmark Dataset for Computer-aided COVID-19 Screening from Chest CT Images
ABSTRACT: Computed tomography (CT) has been widely explored as a COVID-19 screening and
assessment tool to complement RT-PCR testing. To assist radiologists with
CT-based COVID-19 screening, a number of computer-aided systems have been
proposed. However, many proposed systems are built using CT data which is
limited in both quantity and diversity. Motivated to support efforts in the
development of machine learning-driven screening systems, we introduce COVIDx
CT-3, a large-scale multinational benchmark dataset for detection of COVID-19
cases from chest CT images. COVIDx CT-3 includes 431,205 CT slices from 6,068
patients across at least 17 countries, which to the best of our knowledge
represents the largest, most diverse dataset of COVID-19 CT images in
open-access form. Additionally, we examine the data diversity and potential
biases of the COVIDx CT-3 dataset, finding that significant geographic and
class imbalances remain despite efforts to curate data from a wide variety of
sources. | {
"abstract": "Computed tomography (CT) has been widely explored as a COVID-19 screening and\nassessment tool to complement RT-PCR testing. To assist radiologists with\nCT-based COVID-19 screening, a number of computer-aided systems have been\nproposed. However, many proposed systems are built using CT data which is\nlimited in both quantity and diversity. Motivated to support efforts in the\ndevelopment of machine learning-driven screening systems, we introduce COVIDx\nCT-3, a large-scale multinational benchmark dataset for detection of COVID-19\ncases from chest CT images. COVIDx CT-3 includes 431,205 CT slices from 6,068\npatients across at least 17 countries, which to the best of our knowledge\nrepresents the largest, most diverse dataset of COVID-19 CT images in\nopen-access form. Additionally, we examine the data diversity and potential\nbiases of the COVIDx CT-3 dataset, finding that significant geographic and\nclass imbalances remain despite efforts to curate data from a wide variety of\nsources.",
"title": "COVIDx CT-3: A Large-scale, Multinational, Open-Source Benchmark Dataset for Computer-aided COVID-19 Screening from Chest CT Images",
"url": "http://arxiv.org/abs/2206.03043v3"
} | null | null | new_dataset | admin | null | false | null | 14f90c95-ff42-402d-8bb3-ccd3ece8cc0d | null | Validated | {
"text_length": 1157
} | 0new_dataset
|
TITLE: Shifts 2.0: Extending The Dataset of Real Distributional Shifts
ABSTRACT: Distributional shift, or the mismatch between training and deployment data,
is a significant obstacle to the usage of machine learning in high-stakes
industrial applications, such as autonomous driving and medicine. This creates
a need to be able to assess how robustly ML models generalize as well as the
quality of their uncertainty estimates. Standard ML baseline datasets do not
allow these properties to be assessed, as the training, validation and test
data are often identically distributed. Recently, a range of dedicated
benchmarks have appeared, featuring both distributionally matched and shifted
data. Among these benchmarks, the Shifts dataset stands out in terms of the
diversity of tasks as well as the data modalities it features. While most of
the benchmarks are heavily dominated by 2D image classification tasks, Shifts
contains tabular weather forecasting, machine translation, and vehicle motion
prediction tasks. This enables the robustness properties of models to be
assessed on a diverse set of industrial-scale tasks and either universal or
directly applicable task-specific conclusions to be reached. In this paper, we
extend the Shifts Dataset with two datasets sourced from industrial, high-risk
applications of high societal importance. Specifically, we consider the tasks
of segmentation of white matter Multiple Sclerosis lesions in 3D magnetic
resonance brain images and the estimation of power consumption in marine cargo
vessels. Both tasks feature ubiquitous distributional shifts and a strict
safety requirement due to the high cost of errors. These new datasets will
allow researchers to further explore robust generalization and uncertainty
estimation in new situations. In this work, we provide a description of the
dataset and baseline results for both tasks. | {
"abstract": "Distributional shift, or the mismatch between training and deployment data,\nis a significant obstacle to the usage of machine learning in high-stakes\nindustrial applications, such as autonomous driving and medicine. This creates\na need to be able to assess how robustly ML models generalize as well as the\nquality of their uncertainty estimates. Standard ML baseline datasets do not\nallow these properties to be assessed, as the training, validation and test\ndata are often identically distributed. Recently, a range of dedicated\nbenchmarks have appeared, featuring both distributionally matched and shifted\ndata. Among these benchmarks, the Shifts dataset stands out in terms of the\ndiversity of tasks as well as the data modalities it features. While most of\nthe benchmarks are heavily dominated by 2D image classification tasks, Shifts\ncontains tabular weather forecasting, machine translation, and vehicle motion\nprediction tasks. This enables the robustness properties of models to be\nassessed on a diverse set of industrial-scale tasks and either universal or\ndirectly applicable task-specific conclusions to be reached. In this paper, we\nextend the Shifts Dataset with two datasets sourced from industrial, high-risk\napplications of high societal importance. Specifically, we consider the tasks\nof segmentation of white matter Multiple Sclerosis lesions in 3D magnetic\nresonance brain images and the estimation of power consumption in marine cargo\nvessels. Both tasks feature ubiquitous distributional shifts and a strict\nsafety requirement due to the high cost of errors. These new datasets will\nallow researchers to further explore robust generalization and uncertainty\nestimation in new situations. In this work, we provide a description of the\ndataset and baseline results for both tasks.",
"title": "Shifts 2.0: Extending The Dataset of Real Distributional Shifts",
"url": "http://arxiv.org/abs/2206.15407v2"
} | null | null | new_dataset | admin | null | false | null | bb2169db-72b3-4ce2-85ae-d66da1e2fd03 | null | Validated | {
"text_length": 1897
} | 0new_dataset
|
TITLE: Machine Learning Performance Analysis to Predict Stroke Based on Imbalanced Medical Dataset
ABSTRACT: Cerebral stroke, the second most substantial cause of death universally, has
been a primary public health concern over the last few years. With the help of
machine learning techniques, early detection of various stroke alerts is
accessible, which can efficiently prevent or diminish the stroke. Medical
dataset, however, are frequently unbalanced in their class label, with a
tendency to poorly predict minority classes. In this paper, the potential risk
factors for stroke are investigated. Moreover, four distinctive approaches are
applied to improve the classification of the minority class in the imbalanced
stroke dataset, which are the ensemble weight voting classifier, the Synthetic
Minority Over-sampling Technique (SMOTE), Principal Component Analysis with
K-Means Clustering (PCA-Kmeans), Focal Loss with the Deep Neural Network (DNN)
and compare their performance. Through the analysis results, SMOTE and
PCA-Kmeans with DNN-Focal Loss work best for the limited size of a large severe
imbalanced dataset,which is 2-4 times outperform Kaggle work. | {
"abstract": "Cerebral stroke, the second most substantial cause of death universally, has\nbeen a primary public health concern over the last few years. With the help of\nmachine learning techniques, early detection of various stroke alerts is\naccessible, which can efficiently prevent or diminish the stroke. Medical\ndataset, however, are frequently unbalanced in their class label, with a\ntendency to poorly predict minority classes. In this paper, the potential risk\nfactors for stroke are investigated. Moreover, four distinctive approaches are\napplied to improve the classification of the minority class in the imbalanced\nstroke dataset, which are the ensemble weight voting classifier, the Synthetic\nMinority Over-sampling Technique (SMOTE), Principal Component Analysis with\nK-Means Clustering (PCA-Kmeans), Focal Loss with the Deep Neural Network (DNN)\nand compare their performance. Through the analysis results, SMOTE and\nPCA-Kmeans with DNN-Focal Loss work best for the limited size of a large severe\nimbalanced dataset,which is 2-4 times outperform Kaggle work.",
"title": "Machine Learning Performance Analysis to Predict Stroke Based on Imbalanced Medical Dataset",
"url": "http://arxiv.org/abs/2211.07652v1"
} | null | null | no_new_dataset | admin | null | false | null | 935e37df-adc5-49be-835a-24ac558f5c98 | null | Validated | {
"text_length": 1184
} | 1no_new_dataset
|
TITLE: Evaluating resampling methods on a real-life highly imbalanced online credit card payments dataset
ABSTRACT: Various problems of any credit card fraud detection based on machine learning
come from the imbalanced aspect of transaction datasets. Indeed, the number of
frauds compared to the number of regular transactions is tiny and has been
shown to damage learning performances, e.g., at worst, the algorithm can learn
to classify all the transactions as regular. Resampling methods and
cost-sensitive approaches are known to be good candidates to leverage this
issue of imbalanced datasets. This paper evaluates numerous state-of-the-art
resampling methods on a large real-life online credit card payments dataset. We
show they are inefficient because methods are intractable or because metrics do
not exhibit substantial improvements. Our work contributes to this domain in
(1) that we compare many state-of-the-art resampling methods on a large-scale
dataset and in (2) that we use a real-life online credit card payments dataset. | {
"abstract": "Various problems of any credit card fraud detection based on machine learning\ncome from the imbalanced aspect of transaction datasets. Indeed, the number of\nfrauds compared to the number of regular transactions is tiny and has been\nshown to damage learning performances, e.g., at worst, the algorithm can learn\nto classify all the transactions as regular. Resampling methods and\ncost-sensitive approaches are known to be good candidates to leverage this\nissue of imbalanced datasets. This paper evaluates numerous state-of-the-art\nresampling methods on a large real-life online credit card payments dataset. We\nshow they are inefficient because methods are intractable or because metrics do\nnot exhibit substantial improvements. Our work contributes to this domain in\n(1) that we compare many state-of-the-art resampling methods on a large-scale\ndataset and in (2) that we use a real-life online credit card payments dataset.",
"title": "Evaluating resampling methods on a real-life highly imbalanced online credit card payments dataset",
"url": "http://arxiv.org/abs/2206.13152v1"
} | null | null | no_new_dataset | admin | null | false | null | e32c31dc-e430-4cce-bda1-78b8ad39277f | null | Default | {
"text_length": 1058
} | 1no_new_dataset
|
TITLE: The Conversational Short-phrase Speaker Diarization (CSSD) Task: Dataset, Evaluation Metric and Baselines
ABSTRACT: The conversation scenario is one of the most important and most challenging
scenarios for speech processing technologies because people in conversation
respond to each other in a casual style. Detecting the speech activities of
each person in a conversation is vital to downstream tasks, like natural
language processing, machine translation, etc. People refer to the detection
technology of "who speak when" as speaker diarization (SD). Traditionally,
diarization error rate (DER) has been used as the standard evaluation metric of
SD systems for a long time. However, DER fails to give enough importance to
short conversational phrases, which are short but important on the semantic
level. Also, a carefully and accurately manually-annotated testing dataset
suitable for evaluating the conversational SD technologies is still unavailable
in the speech community. In this paper, we design and describe the
Conversational Short-phrases Speaker Diarization (CSSD) task, which consists of
training and testing datasets, evaluation metric and baselines. In the dataset
aspect, despite the previously open-sourced 180-hour conversational
MagicData-RAMC dataset, we prepare an individual 20-hour conversational speech
test dataset with carefully and artificially verified speakers timestamps
annotations for the CSSD task. In the metric aspect, we design the new
conversational DER (CDER) evaluation metric, which calculates the SD accuracy
at the utterance level. In the baseline aspect, we adopt a commonly used
method: Variational Bayes HMM x-vector system, as the baseline of the CSSD
task. Our evaluation metric is publicly available at
https://github.com/SpeechClub/CDER_Metric. | {
"abstract": "The conversation scenario is one of the most important and most challenging\nscenarios for speech processing technologies because people in conversation\nrespond to each other in a casual style. Detecting the speech activities of\neach person in a conversation is vital to downstream tasks, like natural\nlanguage processing, machine translation, etc. People refer to the detection\ntechnology of \"who speak when\" as speaker diarization (SD). Traditionally,\ndiarization error rate (DER) has been used as the standard evaluation metric of\nSD systems for a long time. However, DER fails to give enough importance to\nshort conversational phrases, which are short but important on the semantic\nlevel. Also, a carefully and accurately manually-annotated testing dataset\nsuitable for evaluating the conversational SD technologies is still unavailable\nin the speech community. In this paper, we design and describe the\nConversational Short-phrases Speaker Diarization (CSSD) task, which consists of\ntraining and testing datasets, evaluation metric and baselines. In the dataset\naspect, despite the previously open-sourced 180-hour conversational\nMagicData-RAMC dataset, we prepare an individual 20-hour conversational speech\ntest dataset with carefully and artificially verified speakers timestamps\nannotations for the CSSD task. In the metric aspect, we design the new\nconversational DER (CDER) evaluation metric, which calculates the SD accuracy\nat the utterance level. In the baseline aspect, we adopt a commonly used\nmethod: Variational Bayes HMM x-vector system, as the baseline of the CSSD\ntask. Our evaluation metric is publicly available at\nhttps://github.com/SpeechClub/CDER_Metric.",
"title": "The Conversational Short-phrase Speaker Diarization (CSSD) Task: Dataset, Evaluation Metric and Baselines",
"url": "http://arxiv.org/abs/2208.08042v1"
} | null | null | new_dataset | admin | null | false | null | fd49659a-52a0-4e86-a6b0-5c2555520522 | null | Validated | {
"text_length": 1819
} | 0new_dataset
|
TITLE: Solar Active Region Magnetogram Image Dataset for Studies of Space Weather
ABSTRACT: In this dataset we provide a comprehensive collection of magnetograms (images
quantifying the strength of the magnetic field) from the National Aeronautics
and Space Administration's (NASA's) Solar Dynamics Observatory (SDO). The
dataset incorporates data from three sources and provides SDO Helioseismic and
Magnetic Imager (HMI) magnetograms of solar active regions (regions of large
magnetic flux, generally the source of eruptive events) as well as labels of
corresponding flaring activity. This dataset will be useful for image analysis
or solar physics research related to magnetic structure, its evolution over
time, and its relation to solar flares. The dataset will be of interest to
those researchers investigating automated solar flare prediction methods,
including supervised and unsupervised machine learning (classical and deep),
binary and multi-class classification, and regression. This dataset is a
minimally processed, user configurable dataset of consistently sized images of
solar active regions that can serve as a benchmark dataset for solar flare
prediction research. | {
"abstract": "In this dataset we provide a comprehensive collection of magnetograms (images\nquantifying the strength of the magnetic field) from the National Aeronautics\nand Space Administration's (NASA's) Solar Dynamics Observatory (SDO). The\ndataset incorporates data from three sources and provides SDO Helioseismic and\nMagnetic Imager (HMI) magnetograms of solar active regions (regions of large\nmagnetic flux, generally the source of eruptive events) as well as labels of\ncorresponding flaring activity. This dataset will be useful for image analysis\nor solar physics research related to magnetic structure, its evolution over\ntime, and its relation to solar flares. The dataset will be of interest to\nthose researchers investigating automated solar flare prediction methods,\nincluding supervised and unsupervised machine learning (classical and deep),\nbinary and multi-class classification, and regression. This dataset is a\nminimally processed, user configurable dataset of consistently sized images of\nsolar active regions that can serve as a benchmark dataset for solar flare\nprediction research.",
"title": "Solar Active Region Magnetogram Image Dataset for Studies of Space Weather",
"url": "http://arxiv.org/abs/2305.09492v2"
} | null | null | new_dataset | admin | null | false | null | eb0b9191-def7-4bf5-b92f-74c249692b45 | null | Validated | {
"text_length": 1200
} | 0new_dataset
|
TITLE: Commander's Intent: A Dataset and Modeling Approach for Human-AI Task Specification in Strategic Play
ABSTRACT: Effective Human-AI teaming requires the ability to communicate the goals of
the team and constraints under which you need the agent to operate. Providing
the ability to specify the shared intent or operation criteria of the team can
enable an AI agent to perform its primary function while still being able to
cater to the specific desires of the current team. While significant work has
been conducted to instruct an agent to perform a task, via language or
demonstrations, prior work lacks a focus on building agents which can operate
within the parameters specified by a team. Worse yet, there is a dearth of
research pertaining to enabling humans to provide their specifications through
unstructured, naturalist language. In this paper, we propose the use of goals
and constraints as a scaffold to modulate and evaluate autonomous agents. We
contribute to this field by presenting a novel dataset, and an associated data
collection protocol, which maps language descriptions to goals and constraints
corresponding to specific strategies developed by human participants for the
board game Risk. Leveraging state-of-the-art language models and augmentation
procedures, we develop a machine learning framework which can be used to
identify goals and constraints from unstructured strategy descriptions. To
empirically validate our approach we conduct a human-subjects study to
establish a human-baseline for our dataset. Our results show that our machine
learning architecture is better able to interpret unstructured language
descriptions into strategy specifications than human raters tasked with
performing the same machine translation task (F(1,272.53) = 17.025, p < 0.001). | {
"abstract": "Effective Human-AI teaming requires the ability to communicate the goals of\nthe team and constraints under which you need the agent to operate. Providing\nthe ability to specify the shared intent or operation criteria of the team can\nenable an AI agent to perform its primary function while still being able to\ncater to the specific desires of the current team. While significant work has\nbeen conducted to instruct an agent to perform a task, via language or\ndemonstrations, prior work lacks a focus on building agents which can operate\nwithin the parameters specified by a team. Worse yet, there is a dearth of\nresearch pertaining to enabling humans to provide their specifications through\nunstructured, naturalist language. In this paper, we propose the use of goals\nand constraints as a scaffold to modulate and evaluate autonomous agents. We\ncontribute to this field by presenting a novel dataset, and an associated data\ncollection protocol, which maps language descriptions to goals and constraints\ncorresponding to specific strategies developed by human participants for the\nboard game Risk. Leveraging state-of-the-art language models and augmentation\nprocedures, we develop a machine learning framework which can be used to\nidentify goals and constraints from unstructured strategy descriptions. To\nempirically validate our approach we conduct a human-subjects study to\nestablish a human-baseline for our dataset. Our results show that our machine\nlearning architecture is better able to interpret unstructured language\ndescriptions into strategy specifications than human raters tasked with\nperforming the same machine translation task (F(1,272.53) = 17.025, p < 0.001).",
"title": "Commander's Intent: A Dataset and Modeling Approach for Human-AI Task Specification in Strategic Play",
"url": "http://arxiv.org/abs/2208.08374v1"
} | null | null | new_dataset | admin | null | false | null | f59375fa-83ae-406a-b541-3df9881bafce | null | Validated | {
"text_length": 1815
} | 0new_dataset
|
TITLE: TruEyes: Utilizing Microtasks in Mobile Apps for Crowdsourced Labeling of Machine Learning Datasets
ABSTRACT: The growing use of supervised machine learning in research and industry has
increased the need for labeled datasets. Crowdsourcing has emerged as a popular
method to create data labels. However, working on large batches of tasks leads
to worker fatigue, negatively impacting labeling quality. To address this, we
present TruEyes, a collaborative crowdsourcing system, enabling the
distribution of micro-tasks to mobile app users. TruEyes allows machine
learning practitioners to publish labeling tasks, mobile app developers to
integrate task ads for monetization, and users to label data instead of
watching advertisements. To evaluate the system, we conducted an experiment
with N=296 participants. Our results show that the quality of the labeled data
is comparable to traditional crowdsourcing approaches and most users prefer
task ads over traditional ads. We discuss extensions to the system and address
how mobile advertisement space can be used as a productive resource in the
future. | {
"abstract": "The growing use of supervised machine learning in research and industry has\nincreased the need for labeled datasets. Crowdsourcing has emerged as a popular\nmethod to create data labels. However, working on large batches of tasks leads\nto worker fatigue, negatively impacting labeling quality. To address this, we\npresent TruEyes, a collaborative crowdsourcing system, enabling the\ndistribution of micro-tasks to mobile app users. TruEyes allows machine\nlearning practitioners to publish labeling tasks, mobile app developers to\nintegrate task ads for monetization, and users to label data instead of\nwatching advertisements. To evaluate the system, we conducted an experiment\nwith N=296 participants. Our results show that the quality of the labeled data\nis comparable to traditional crowdsourcing approaches and most users prefer\ntask ads over traditional ads. We discuss extensions to the system and address\nhow mobile advertisement space can be used as a productive resource in the\nfuture.",
"title": "TruEyes: Utilizing Microtasks in Mobile Apps for Crowdsourced Labeling of Machine Learning Datasets",
"url": "http://arxiv.org/abs/2209.14708v1"
} | null | null | no_new_dataset | admin | null | false | null | 72f7f796-bceb-44cc-a1c2-93ddbc5e027c | null | Validated | {
"text_length": 1126
} | 1no_new_dataset
|
TITLE: Data Models for Dataset Drift Controls in Machine Learning With Optical Images
ABSTRACT: Camera images are ubiquitous in machine learning research. They also play a
central role in the delivery of important services spanning medicine and
environmental surveying. However, the application of machine learning models in
these domains has been limited because of robustness concerns. A primary
failure mode are performance drops due to differences between the training and
deployment data. While there are methods to prospectively validate the
robustness of machine learning models to such dataset drifts, existing
approaches do not account for explicit models of the primary object of
interest: the data. This limits our ability to study and understand the
relationship between data generation and downstream machine learning model
performance in a physically accurate manner. In this study, we demonstrate how
to overcome this limitation by pairing traditional machine learning with
physical optics to obtain explicit and differentiable data models. We
demonstrate how such data models can be constructed for image data and used to
control downstream machine learning model performance related to dataset drift.
The findings are distilled into three applications. First, drift synthesis
enables the controlled generation of physically faithful drift test cases to
power model selection and targeted generalization. Second, the gradient
connection between machine learning task model and data model allows advanced,
precise tolerancing of task model sensitivity to changes in the data
generation. These drift forensics can be used to precisely specify the
acceptable data environments in which a task model may be run. Third, drift
optimization opens up the possibility to create drifts that can help the task
model learn better faster, effectively optimizing the data generating process
itself. A guide to access the open code and datasets is available at
https://github.com/aiaudit-org/raw2logit. | {
"abstract": "Camera images are ubiquitous in machine learning research. They also play a\ncentral role in the delivery of important services spanning medicine and\nenvironmental surveying. However, the application of machine learning models in\nthese domains has been limited because of robustness concerns. A primary\nfailure mode are performance drops due to differences between the training and\ndeployment data. While there are methods to prospectively validate the\nrobustness of machine learning models to such dataset drifts, existing\napproaches do not account for explicit models of the primary object of\ninterest: the data. This limits our ability to study and understand the\nrelationship between data generation and downstream machine learning model\nperformance in a physically accurate manner. In this study, we demonstrate how\nto overcome this limitation by pairing traditional machine learning with\nphysical optics to obtain explicit and differentiable data models. We\ndemonstrate how such data models can be constructed for image data and used to\ncontrol downstream machine learning model performance related to dataset drift.\nThe findings are distilled into three applications. First, drift synthesis\nenables the controlled generation of physically faithful drift test cases to\npower model selection and targeted generalization. Second, the gradient\nconnection between machine learning task model and data model allows advanced,\nprecise tolerancing of task model sensitivity to changes in the data\ngeneration. These drift forensics can be used to precisely specify the\nacceptable data environments in which a task model may be run. Third, drift\noptimization opens up the possibility to create drifts that can help the task\nmodel learn better faster, effectively optimizing the data generating process\nitself. A guide to access the open code and datasets is available at\nhttps://github.com/aiaudit-org/raw2logit.",
"title": "Data Models for Dataset Drift Controls in Machine Learning With Optical Images",
"url": "http://arxiv.org/abs/2211.02578v3"
} | null | null | no_new_dataset | admin | null | false | null | 75bfa54b-db90-45f0-89a9-3102ff3b3df3 | null | Validated | {
"text_length": 2020
} | 1no_new_dataset
|
TITLE: The ITU Faroese Pairs Dataset
ABSTRACT: This article documents a dataset of sentence pairs between Faroese and
Danish, produced at ITU Copenhagen. The data covers tranlsation from both
source languages, and is intended for use as training data for machine
translation systems in this language pair. | {
"abstract": "This article documents a dataset of sentence pairs between Faroese and\nDanish, produced at ITU Copenhagen. The data covers tranlsation from both\nsource languages, and is intended for use as training data for machine\ntranslation systems in this language pair.",
"title": "The ITU Faroese Pairs Dataset",
"url": "http://arxiv.org/abs/2206.08727v1"
} | null | null | new_dataset | admin | null | false | null | 2f29679a-049c-4cee-b4d7-97905b3c9b6c | null | Validated | {
"text_length": 322
} | 0new_dataset
|
TITLE: Improving Multilayer-Perceptron(MLP)-based Network Anomaly Detection with Birch Clustering on CICIDS-2017 Dataset
ABSTRACT: Machine learning algorithms have been widely used in intrusion detection
systems, including Multi-layer Perceptron (MLP). In this study, we proposed a
two-stage model that combines the Birch clustering algorithm and MLP classifier
to improve the performance of network anomaly multi-classification. In our
proposed method, we first apply Birch or Kmeans as an unsupervised clustering
algorithm to the CICIDS-2017 dataset to pre-group the data. The generated
pseudo-label is then added as an additional feature to the training of the
MLP-based classifier. The experimental results show that using Birch and
K-Means clustering for data pre-grouping can improve intrusion detection system
performance. Our method can achieve 99.73% accuracy in multi-classification
using Birch clustering, which is better than similar researches using a
stand-alone MLP model. | {
"abstract": "Machine learning algorithms have been widely used in intrusion detection\nsystems, including Multi-layer Perceptron (MLP). In this study, we proposed a\ntwo-stage model that combines the Birch clustering algorithm and MLP classifier\nto improve the performance of network anomaly multi-classification. In our\nproposed method, we first apply Birch or Kmeans as an unsupervised clustering\nalgorithm to the CICIDS-2017 dataset to pre-group the data. The generated\npseudo-label is then added as an additional feature to the training of the\nMLP-based classifier. The experimental results show that using Birch and\nK-Means clustering for data pre-grouping can improve intrusion detection system\nperformance. Our method can achieve 99.73% accuracy in multi-classification\nusing Birch clustering, which is better than similar researches using a\nstand-alone MLP model.",
"title": "Improving Multilayer-Perceptron(MLP)-based Network Anomaly Detection with Birch Clustering on CICIDS-2017 Dataset",
"url": "http://arxiv.org/abs/2208.09711v2"
} | null | null | no_new_dataset | admin | null | false | null | 60902729-ac8e-43f0-b661-6bdcc36c3f92 | null | Validated | {
"text_length": 1004
} | 1no_new_dataset
|
TITLE: PcMSP: A Dataset for Scientific Action Graphs Extraction from Polycrystalline Materials Synthesis Procedure Text
ABSTRACT: Scientific action graphs extraction from materials synthesis procedures is
important for reproducible research, machine automation, and material
prediction. But the lack of annotated data has hindered progress in this field.
We demonstrate an effort to annotate Polycrystalline Materials Synthesis
Procedures (PcMSP) from 305 open access scientific articles for the
construction of synthesis action graphs. This is a new dataset for material
science information extraction that simultaneously contains the synthesis
sentences extracted from the experimental paragraphs, as well as the entity
mentions and intra-sentence relations. A two-step human annotation and
inter-annotator agreement study guarantee the high quality of the PcMSP corpus.
We introduce four natural language processing tasks: sentence classification,
named entity recognition, relation classification, and joint extraction of
entities and relations. Comprehensive experiments validate the effectiveness of
several state-of-the-art models for these challenges while leaving large space
for improvement. We also perform the error analysis and point out some unique
challenges that require further investigation. We will release our annotation
scheme, the corpus, and codes to the research community to alleviate the
scarcity of labeled data in this domain. | {
"abstract": "Scientific action graphs extraction from materials synthesis procedures is\nimportant for reproducible research, machine automation, and material\nprediction. But the lack of annotated data has hindered progress in this field.\nWe demonstrate an effort to annotate Polycrystalline Materials Synthesis\nProcedures (PcMSP) from 305 open access scientific articles for the\nconstruction of synthesis action graphs. This is a new dataset for material\nscience information extraction that simultaneously contains the synthesis\nsentences extracted from the experimental paragraphs, as well as the entity\nmentions and intra-sentence relations. A two-step human annotation and\ninter-annotator agreement study guarantee the high quality of the PcMSP corpus.\nWe introduce four natural language processing tasks: sentence classification,\nnamed entity recognition, relation classification, and joint extraction of\nentities and relations. Comprehensive experiments validate the effectiveness of\nseveral state-of-the-art models for these challenges while leaving large space\nfor improvement. We also perform the error analysis and point out some unique\nchallenges that require further investigation. We will release our annotation\nscheme, the corpus, and codes to the research community to alleviate the\nscarcity of labeled data in this domain.",
"title": "PcMSP: A Dataset for Scientific Action Graphs Extraction from Polycrystalline Materials Synthesis Procedure Text",
"url": "http://arxiv.org/abs/2210.12401v1"
} | null | null | new_dataset | admin | null | false | null | 5f81a480-6c41-422d-a3f5-cf0ea5d181b3 | null | Validated | {
"text_length": 1471
} | 0new_dataset
|
TITLE: Semi-Supervised Semantic Segmentation in Earth Observation: The MiniFrance Suite, Dataset Analysis and Multi-task Network Study
ABSTRACT: The development of semi-supervised learning techniques is essential to
enhance the generalization capacities of machine learning algorithms. Indeed,
raw image data are abundant while labels are scarce, therefore it is crucial to
leverage unlabeled inputs to build better models. The availability of large
databases have been key for the development of learning algorithms with high
level performance.
Despite the major role of machine learning in Earth Observation to derive
products such as land cover maps, datasets in the field are still limited,
either because of modest surface coverage, lack of variety of scenes or
restricted classes to identify. We introduce a novel large-scale dataset for
semi-supervised semantic segmentation in Earth Observation, the MiniFrance
suite. MiniFrance has several unprecedented properties: it is large-scale,
containing over 2000 very high resolution aerial images, accounting for more
than 200 billions samples (pixels); it is varied, covering 16 conurbations in
France, with various climates, different landscapes, and urban as well as
countryside scenes; and it is challenging, considering land use classes with
high-level semantics. Nevertheless, the most distinctive quality of MiniFrance
is being the only dataset in the field especially designed for semi-supervised
learning: it contains labeled and unlabeled images in its training partition,
which reproduces a life-like scenario. Along with this dataset, we present
tools for data representativeness analysis in terms of appearance similarity
and a thorough study of MiniFrance data, demonstrating that it is suitable for
learning and generalizes well in a semi-supervised setting. Finally, we present
semi-supervised deep architectures based on multi-task learning and the first
experiments on MiniFrance. | {
"abstract": "The development of semi-supervised learning techniques is essential to\nenhance the generalization capacities of machine learning algorithms. Indeed,\nraw image data are abundant while labels are scarce, therefore it is crucial to\nleverage unlabeled inputs to build better models. The availability of large\ndatabases have been key for the development of learning algorithms with high\nlevel performance.\n Despite the major role of machine learning in Earth Observation to derive\nproducts such as land cover maps, datasets in the field are still limited,\neither because of modest surface coverage, lack of variety of scenes or\nrestricted classes to identify. We introduce a novel large-scale dataset for\nsemi-supervised semantic segmentation in Earth Observation, the MiniFrance\nsuite. MiniFrance has several unprecedented properties: it is large-scale,\ncontaining over 2000 very high resolution aerial images, accounting for more\nthan 200 billions samples (pixels); it is varied, covering 16 conurbations in\nFrance, with various climates, different landscapes, and urban as well as\ncountryside scenes; and it is challenging, considering land use classes with\nhigh-level semantics. Nevertheless, the most distinctive quality of MiniFrance\nis being the only dataset in the field especially designed for semi-supervised\nlearning: it contains labeled and unlabeled images in its training partition,\nwhich reproduces a life-like scenario. Along with this dataset, we present\ntools for data representativeness analysis in terms of appearance similarity\nand a thorough study of MiniFrance data, demonstrating that it is suitable for\nlearning and generalizes well in a semi-supervised setting. Finally, we present\nsemi-supervised deep architectures based on multi-task learning and the first\nexperiments on MiniFrance.",
"title": "Semi-Supervised Semantic Segmentation in Earth Observation: The MiniFrance Suite, Dataset Analysis and Multi-task Network Study",
"url": "http://arxiv.org/abs/2010.07830v1"
} | null | null | new_dataset | admin | null | false | null | 27070983-724c-4a1f-b90c-e1d8738d2816 | null | Validated | {
"text_length": 1970
} | 0new_dataset
|
TITLE: RaidaR: A Rich Annotated Image Dataset of Rainy Street Scenes
ABSTRACT: We introduce RaidaR, a rich annotated image dataset of rainy street scenes,
to support autonomous driving research. The new dataset contains the largest
number of rainy images (58,542) to date, 5,000 of which provide semantic
segmentations and 3,658 provide object instance segmentations. The RaidaR
images cover a wide range of realistic rain-induced artifacts, including fog,
droplets, and road reflections, which can effectively augment existing street
scene datasets to improve data-driven machine perception during rainy weather.
To facilitate efficient annotation of a large volume of images, we develop a
semi-automatic scheme combining manual segmentation and an automated processing
akin to cross validation, resulting in 10-20 fold reduction on annotation time.
We demonstrate the utility of our new dataset by showing how data augmentation
with RaidaR can elevate the accuracy of existing segmentation algorithms. We
also present a novel unpaired image-to-image translation algorithm for
adding/removing rain artifacts, which directly benefits from RaidaR. | {
"abstract": "We introduce RaidaR, a rich annotated image dataset of rainy street scenes,\nto support autonomous driving research. The new dataset contains the largest\nnumber of rainy images (58,542) to date, 5,000 of which provide semantic\nsegmentations and 3,658 provide object instance segmentations. The RaidaR\nimages cover a wide range of realistic rain-induced artifacts, including fog,\ndroplets, and road reflections, which can effectively augment existing street\nscene datasets to improve data-driven machine perception during rainy weather.\nTo facilitate efficient annotation of a large volume of images, we develop a\nsemi-automatic scheme combining manual segmentation and an automated processing\nakin to cross validation, resulting in 10-20 fold reduction on annotation time.\nWe demonstrate the utility of our new dataset by showing how data augmentation\nwith RaidaR can elevate the accuracy of existing segmentation algorithms. We\nalso present a novel unpaired image-to-image translation algorithm for\nadding/removing rain artifacts, which directly benefits from RaidaR.",
"title": "RaidaR: A Rich Annotated Image Dataset of Rainy Street Scenes",
"url": "http://arxiv.org/abs/2104.04606v3"
} | null | null | new_dataset | admin | null | false | null | 2811a11c-72ae-43e1-bf62-d086501ece10 | null | Validated | {
"text_length": 1163
} | 0new_dataset
|
TITLE: AIR-Act2Act: Human-human interaction dataset for teaching non-verbal social behaviors to robots
ABSTRACT: To better interact with users, a social robot should understand the users'
behavior, infer the intention, and respond appropriately. Machine learning is
one way of implementing robot intelligence. It provides the ability to
automatically learn and improve from experience instead of explicitly telling
the robot what to do. Social skills can also be learned through watching
human-human interaction videos. However, human-human interaction datasets are
relatively scarce to learn interactions that occur in various situations.
Moreover, we aim to use service robots in the elderly-care domain; however,
there has been no interaction dataset collected for this domain. For this
reason, we introduce a human-human interaction dataset for teaching non-verbal
social behaviors to robots. It is the only interaction dataset that elderly
people have participated in as performers. We recruited 100 elderly people and
two college students to perform 10 interactions in an indoor environment. The
entire dataset has 5,000 interaction samples, each of which contains depth
maps, body indexes and 3D skeletal data that are captured with three Microsoft
Kinect v2 cameras. In addition, we provide the joint angles of a humanoid NAO
robot which are converted from the human behavior that robots need to learn.
The dataset and useful python scripts are available for download at
https://github.com/ai4r/AIR-Act2Act. It can be used to not only teach social
skills to robots but also benchmark action recognition algorithms. | {
"abstract": "To better interact with users, a social robot should understand the users'\nbehavior, infer the intention, and respond appropriately. Machine learning is\none way of implementing robot intelligence. It provides the ability to\nautomatically learn and improve from experience instead of explicitly telling\nthe robot what to do. Social skills can also be learned through watching\nhuman-human interaction videos. However, human-human interaction datasets are\nrelatively scarce to learn interactions that occur in various situations.\nMoreover, we aim to use service robots in the elderly-care domain; however,\nthere has been no interaction dataset collected for this domain. For this\nreason, we introduce a human-human interaction dataset for teaching non-verbal\nsocial behaviors to robots. It is the only interaction dataset that elderly\npeople have participated in as performers. We recruited 100 elderly people and\ntwo college students to perform 10 interactions in an indoor environment. The\nentire dataset has 5,000 interaction samples, each of which contains depth\nmaps, body indexes and 3D skeletal data that are captured with three Microsoft\nKinect v2 cameras. In addition, we provide the joint angles of a humanoid NAO\nrobot which are converted from the human behavior that robots need to learn.\nThe dataset and useful python scripts are available for download at\nhttps://github.com/ai4r/AIR-Act2Act. It can be used to not only teach social\nskills to robots but also benchmark action recognition algorithms.",
"title": "AIR-Act2Act: Human-human interaction dataset for teaching non-verbal social behaviors to robots",
"url": "http://arxiv.org/abs/2009.02041v1"
} | null | null | new_dataset | admin | null | false | null | 0aa0a75f-ff3b-4620-9932-64ad6aea89e4 | null | Validated | {
"text_length": 1639
} | 0new_dataset
|
TITLE: Interpreting Black-box Machine Learning Models for High Dimensional Datasets
ABSTRACT: Deep neural networks (DNNs) have been shown to outperform traditional machine
learning algorithms in a broad variety of application domains due to their
effectiveness in modeling complex problems and handling high-dimensional
datasets. Many real-life datasets, however, are of increasingly high
dimensionality, where a large number of features may be irrelevant for both
supervised and unsupervised learning tasks. The inclusion of such features
would not only introduce unwanted noise but also increase computational
complexity. Furthermore, due to high non-linearity and dependency among a large
number of features, DNN models tend to be unavoidably opaque and perceived as
black-box methods because of their not well-understood internal functioning.
Their algorithmic complexity is often simply beyond the capacities of humans to
understand the interplay among myriads of hyperparameters. A well-interpretable
model can identify statistically significant features and explain the way they
affect the model's outcome. In this paper, we propose an efficient method to
improve the interpretability of black-box models for classification tasks in
the case of high-dimensional datasets. First, we train a black-box model on a
high-dimensional dataset to learn the embeddings on which the classification is
performed. To decompose the inner working principles of the black-box model and
to identify top-k important features, we employ different probing and
perturbing techniques. We then approximate the behavior of the black-box model
by means of an interpretable surrogate model on the top-k feature space.
Finally, we derive decision rules and local explanations from the surrogate
model to explain individual decisions. Our approach outperforms
state-of-the-art methods like TabNet and XGboost when tested on different
datasets with varying dimensionality between 50 and 20,000 w.r.t metrics and
explainability. | {
"abstract": "Deep neural networks (DNNs) have been shown to outperform traditional machine\nlearning algorithms in a broad variety of application domains due to their\neffectiveness in modeling complex problems and handling high-dimensional\ndatasets. Many real-life datasets, however, are of increasingly high\ndimensionality, where a large number of features may be irrelevant for both\nsupervised and unsupervised learning tasks. The inclusion of such features\nwould not only introduce unwanted noise but also increase computational\ncomplexity. Furthermore, due to high non-linearity and dependency among a large\nnumber of features, DNN models tend to be unavoidably opaque and perceived as\nblack-box methods because of their not well-understood internal functioning.\nTheir algorithmic complexity is often simply beyond the capacities of humans to\nunderstand the interplay among myriads of hyperparameters. A well-interpretable\nmodel can identify statistically significant features and explain the way they\naffect the model's outcome. In this paper, we propose an efficient method to\nimprove the interpretability of black-box models for classification tasks in\nthe case of high-dimensional datasets. First, we train a black-box model on a\nhigh-dimensional dataset to learn the embeddings on which the classification is\nperformed. To decompose the inner working principles of the black-box model and\nto identify top-k important features, we employ different probing and\nperturbing techniques. We then approximate the behavior of the black-box model\nby means of an interpretable surrogate model on the top-k feature space.\nFinally, we derive decision rules and local explanations from the surrogate\nmodel to explain individual decisions. Our approach outperforms\nstate-of-the-art methods like TabNet and XGboost when tested on different\ndatasets with varying dimensionality between 50 and 20,000 w.r.t metrics and\nexplainability.",
"title": "Interpreting Black-box Machine Learning Models for High Dimensional Datasets",
"url": "http://arxiv.org/abs/2208.13405v2"
} | null | null | no_new_dataset | admin | null | false | null | a7139c48-a50b-4a25-b653-75cfe55224fe | null | Validated | {
"text_length": 2023
} | 1no_new_dataset
|
TITLE: XLCoST: A Benchmark Dataset for Cross-lingual Code Intelligence
ABSTRACT: Recent advances in machine learning have significantly improved the
understanding of source code data and achieved good performance on a number of
downstream tasks. Open source repositories like GitHub enable this process with
rich unlabeled code data. However, the lack of high quality labeled data has
largely hindered the progress of several code related tasks, such as program
translation, summarization, synthesis, and code search. This paper introduces
XLCoST, Cross-Lingual Code SnippeT dataset, a new benchmark dataset for
cross-lingual code intelligence. Our dataset contains fine-grained parallel
data from 8 languages (7 commonly used programming languages and English), and
supports 10 cross-lingual code tasks. To the best of our knowledge, it is the
largest parallel dataset for source code both in terms of size and the number
of languages. We also provide the performance of several state-of-the-art
baseline models for each task. We believe this new dataset can be a valuable
asset for the research community and facilitate the development and validation
of new methods for cross-lingual code intelligence. | {
"abstract": "Recent advances in machine learning have significantly improved the\nunderstanding of source code data and achieved good performance on a number of\ndownstream tasks. Open source repositories like GitHub enable this process with\nrich unlabeled code data. However, the lack of high quality labeled data has\nlargely hindered the progress of several code related tasks, such as program\ntranslation, summarization, synthesis, and code search. This paper introduces\nXLCoST, Cross-Lingual Code SnippeT dataset, a new benchmark dataset for\ncross-lingual code intelligence. Our dataset contains fine-grained parallel\ndata from 8 languages (7 commonly used programming languages and English), and\nsupports 10 cross-lingual code tasks. To the best of our knowledge, it is the\nlargest parallel dataset for source code both in terms of size and the number\nof languages. We also provide the performance of several state-of-the-art\nbaseline models for each task. We believe this new dataset can be a valuable\nasset for the research community and facilitate the development and validation\nof new methods for cross-lingual code intelligence.",
"title": "XLCoST: A Benchmark Dataset for Cross-lingual Code Intelligence",
"url": "http://arxiv.org/abs/2206.08474v1"
} | null | null | new_dataset | admin | null | false | null | a105b929-87c4-4981-99c6-4d6687c61b36 | null | Validated | {
"text_length": 1221
} | 0new_dataset
|
TITLE: IFCNet: A Benchmark Dataset for IFC Entity Classification
ABSTRACT: Enhancing interoperability and information exchange between domain-specific
software products for BIM is an important aspect in the Architecture,
Engineering, Construction and Operations industry. Recent research started
investigating methods from the areas of machine and deep learning for semantic
enrichment of BIM models. However, training and evaluation of these machine
learning algorithms requires sufficiently large and comprehensive datasets.
This work presents IFCNet, a dataset of single-entity IFC files spanning a
broad range of IFC classes containing both geometric and semantic information.
Using only the geometric information of objects, the experiments show that
three different deep learning models are able to achieve good classification
performance. | {
"abstract": "Enhancing interoperability and information exchange between domain-specific\nsoftware products for BIM is an important aspect in the Architecture,\nEngineering, Construction and Operations industry. Recent research started\ninvestigating methods from the areas of machine and deep learning for semantic\nenrichment of BIM models. However, training and evaluation of these machine\nlearning algorithms requires sufficiently large and comprehensive datasets.\nThis work presents IFCNet, a dataset of single-entity IFC files spanning a\nbroad range of IFC classes containing both geometric and semantic information.\nUsing only the geometric information of objects, the experiments show that\nthree different deep learning models are able to achieve good classification\nperformance.",
"title": "IFCNet: A Benchmark Dataset for IFC Entity Classification",
"url": "http://arxiv.org/abs/2106.09712v1"
} | null | null | new_dataset | admin | null | false | null | e6fb60d6-17c4-4d32-a8b2-8666742d1ee0 | null | Validated | {
"text_length": 862
} | 0new_dataset
|
TITLE: Intrusion Detection Systems Using Support Vector Machines on the KDDCUP'99 and NSL-KDD Datasets: A Comprehensive Survey
ABSTRACT: With the growing rates of cyber-attacks and cyber espionage, the need for
better and more powerful intrusion detection systems (IDS) is even more
warranted nowadays. The basic task of an IDS is to act as the first line of
defense, in detecting attacks on the internet. As intrusion tactics from
intruders become more sophisticated and difficult to detect, researchers have
started to apply novel Machine Learning (ML) techniques to effectively detect
intruders and hence preserve internet users' information and overall trust in
the entire internet network security. Over the last decade, there has been an
explosion of research on intrusion detection techniques based on ML and Deep
Learning (DL) architectures on various cyber security-based datasets such as
the DARPA, KDDCUP'99, NSL-KDD, CAIDA, CTU-13, UNSW-NB15. In this research, we
review contemporary literature and provide a comprehensive survey of different
types of intrusion detection technique that applies Support Vector Machines
(SVMs) algorithms as a classifier. We focus only on studies that have been
evaluated on the two most widely used datasets in cybersecurity namely: the
KDDCUP'99 and the NSL-KDD datasets. We provide a summary of each method,
identifying the role of the SVMs classifier, and all other algorithms involved
in the studies. Furthermore, we present a critical review of each method, in
tabular form, highlighting the performance measures, strengths, and limitations
of each of the methods surveyed. | {
"abstract": "With the growing rates of cyber-attacks and cyber espionage, the need for\nbetter and more powerful intrusion detection systems (IDS) is even more\nwarranted nowadays. The basic task of an IDS is to act as the first line of\ndefense, in detecting attacks on the internet. As intrusion tactics from\nintruders become more sophisticated and difficult to detect, researchers have\nstarted to apply novel Machine Learning (ML) techniques to effectively detect\nintruders and hence preserve internet users' information and overall trust in\nthe entire internet network security. Over the last decade, there has been an\nexplosion of research on intrusion detection techniques based on ML and Deep\nLearning (DL) architectures on various cyber security-based datasets such as\nthe DARPA, KDDCUP'99, NSL-KDD, CAIDA, CTU-13, UNSW-NB15. In this research, we\nreview contemporary literature and provide a comprehensive survey of different\ntypes of intrusion detection technique that applies Support Vector Machines\n(SVMs) algorithms as a classifier. We focus only on studies that have been\nevaluated on the two most widely used datasets in cybersecurity namely: the\nKDDCUP'99 and the NSL-KDD datasets. We provide a summary of each method,\nidentifying the role of the SVMs classifier, and all other algorithms involved\nin the studies. Furthermore, we present a critical review of each method, in\ntabular form, highlighting the performance measures, strengths, and limitations\nof each of the methods surveyed.",
"title": "Intrusion Detection Systems Using Support Vector Machines on the KDDCUP'99 and NSL-KDD Datasets: A Comprehensive Survey",
"url": "http://arxiv.org/abs/2209.05579v1"
} | null | null | no_new_dataset | admin | null | false | null | 0b5e25e8-a4cd-4de2-818b-046552e7461f | null | Validated | {
"text_length": 1640
} | 1no_new_dataset
|
TITLE: NEREL-BIO: A Dataset of Biomedical Abstracts Annotated with Nested Named Entities
ABSTRACT: This paper describes NEREL-BIO -- an annotation scheme and corpus of PubMed
abstracts in Russian and smaller number of abstracts in English. NEREL-BIO
extends the general domain dataset NEREL by introducing domain-specific entity
types. NEREL-BIO annotation scheme covers both general and biomedical domains
making it suitable for domain transfer experiments. NEREL-BIO provides
annotation for nested named entities as an extension of the scheme employed for
NEREL. Nested named entities may cross entity boundaries to connect to shorter
entities nested within longer entities, making them harder to detect.
NEREL-BIO contains annotations for 700+ Russian and 100+ English abstracts.
All English PubMed annotations have corresponding Russian counterparts. Thus,
NEREL-BIO comprises the following specific features: annotation of nested named
entities, it can be used as a benchmark for cross-domain (NEREL -> NEREL-BIO)
and cross-language (English -> Russian) transfer. We experiment with both
transformer-based sequence models and machine reading comprehension (MRC)
models and report their results.
The dataset is freely available at https://github.com/nerel-ds/NEREL-BIO. | {
"abstract": "This paper describes NEREL-BIO -- an annotation scheme and corpus of PubMed\nabstracts in Russian and smaller number of abstracts in English. NEREL-BIO\nextends the general domain dataset NEREL by introducing domain-specific entity\ntypes. NEREL-BIO annotation scheme covers both general and biomedical domains\nmaking it suitable for domain transfer experiments. NEREL-BIO provides\nannotation for nested named entities as an extension of the scheme employed for\nNEREL. Nested named entities may cross entity boundaries to connect to shorter\nentities nested within longer entities, making them harder to detect.\n NEREL-BIO contains annotations for 700+ Russian and 100+ English abstracts.\nAll English PubMed annotations have corresponding Russian counterparts. Thus,\nNEREL-BIO comprises the following specific features: annotation of nested named\nentities, it can be used as a benchmark for cross-domain (NEREL -> NEREL-BIO)\nand cross-language (English -> Russian) transfer. We experiment with both\ntransformer-based sequence models and machine reading comprehension (MRC)\nmodels and report their results.\n The dataset is freely available at https://github.com/nerel-ds/NEREL-BIO.",
"title": "NEREL-BIO: A Dataset of Biomedical Abstracts Annotated with Nested Named Entities",
"url": "http://arxiv.org/abs/2210.11913v1"
} | null | null | new_dataset | admin | null | false | null | 17db3f79-0176-40de-8448-c76da8578952 | null | Validated | {
"text_length": 1294
} | 0new_dataset
|
TITLE: Towards emotion recognition for virtual environments: an evaluation of EEG features on benchmark dataset
ABSTRACT: One of the challenges in virtual environments is the difficulty users have in
interacting with these increasingly complex systems. Ultimately, endowing
machines with the ability to perceive users emotions will enable a more
intuitive and reliable interaction. Consequently, using the
electroencephalogram as a bio-signal sensor, the affective state of a user can
be modelled and subsequently utilised in order to achieve a system that can
recognise and react to the user's emotions. This paper investigates features
extracted from electroencephalogram signals for the purpose of affective state
modelling based on Russell's Circumplex Model. Investigations are presented
that aim to provide the foundation for future work in modelling user affect to
enhance interaction experience in virtual environments. The DEAP dataset was
used within this work, along with a Support Vector Machine and Random Forest,
which yielded reasonable classification accuracies for Valence and Arousal
using feature vectors based on statistical measurements and band power from the
\'z, \b{eta}, \'z, and \'z\'z waves and High Order Crossing of the EEG signal. | {
"abstract": "One of the challenges in virtual environments is the difficulty users have in\ninteracting with these increasingly complex systems. Ultimately, endowing\nmachines with the ability to perceive users emotions will enable a more\nintuitive and reliable interaction. Consequently, using the\nelectroencephalogram as a bio-signal sensor, the affective state of a user can\nbe modelled and subsequently utilised in order to achieve a system that can\nrecognise and react to the user's emotions. This paper investigates features\nextracted from electroencephalogram signals for the purpose of affective state\nmodelling based on Russell's Circumplex Model. Investigations are presented\nthat aim to provide the foundation for future work in modelling user affect to\nenhance interaction experience in virtual environments. The DEAP dataset was\nused within this work, along with a Support Vector Machine and Random Forest,\nwhich yielded reasonable classification accuracies for Valence and Arousal\nusing feature vectors based on statistical measurements and band power from the\n\\'z, \\b{eta}, \\'z, and \\'z\\'z waves and High Order Crossing of the EEG signal.",
"title": "Towards emotion recognition for virtual environments: an evaluation of EEG features on benchmark dataset",
"url": "http://arxiv.org/abs/2210.13876v1"
} | null | null | no_new_dataset | admin | null | false | null | e755446b-2f17-4138-8928-166ca1d0359a | null | Validated | {
"text_length": 1277
} | 1no_new_dataset
|
TITLE: METS-CoV: A Dataset of Medical Entity and Targeted Sentiment on COVID-19 Related Tweets
ABSTRACT: The COVID-19 pandemic continues to bring up various topics discussed or
debated on social media. In order to explore the impact of pandemics on
people's lives, it is crucial to understand the public's concerns and attitudes
towards pandemic-related entities (e.g., drugs, vaccines) on social media.
However, models trained on existing named entity recognition (NER) or targeted
sentiment analysis (TSA) datasets have limited ability to understand
COVID-19-related social media texts because these datasets are not designed or
annotated from a medical perspective. This paper releases METS-CoV, a dataset
containing medical entities and targeted sentiments from COVID-19-related
tweets. METS-CoV contains 10,000 tweets with 7 types of entities, including 4
medical entity types (Disease, Drug, Symptom, and Vaccine) and 3 general entity
types (Person, Location, and Organization). To further investigate tweet users'
attitudes toward specific entities, 4 types of entities (Person, Organization,
Drug, and Vaccine) are selected and annotated with user sentiments, resulting
in a targeted sentiment dataset with 9,101 entities (in 5,278 tweets). To the
best of our knowledge, METS-CoV is the first dataset to collect medical
entities and corresponding sentiments of COVID-19-related tweets. We benchmark
the performance of classical machine learning models and state-of-the-art deep
learning models on NER and TSA tasks with extensive experiments. Results show
that the dataset has vast room for improvement for both NER and TSA tasks.
METS-CoV is an important resource for developing better medical social media
tools and facilitating computational social science research, especially in
epidemiology. Our data, annotation guidelines, benchmark models, and source
code are publicly available (https://github.com/YLab-Open/METS-CoV) to ensure
reproducibility. | {
"abstract": "The COVID-19 pandemic continues to bring up various topics discussed or\ndebated on social media. In order to explore the impact of pandemics on\npeople's lives, it is crucial to understand the public's concerns and attitudes\ntowards pandemic-related entities (e.g., drugs, vaccines) on social media.\nHowever, models trained on existing named entity recognition (NER) or targeted\nsentiment analysis (TSA) datasets have limited ability to understand\nCOVID-19-related social media texts because these datasets are not designed or\nannotated from a medical perspective. This paper releases METS-CoV, a dataset\ncontaining medical entities and targeted sentiments from COVID-19-related\ntweets. METS-CoV contains 10,000 tweets with 7 types of entities, including 4\nmedical entity types (Disease, Drug, Symptom, and Vaccine) and 3 general entity\ntypes (Person, Location, and Organization). To further investigate tweet users'\nattitudes toward specific entities, 4 types of entities (Person, Organization,\nDrug, and Vaccine) are selected and annotated with user sentiments, resulting\nin a targeted sentiment dataset with 9,101 entities (in 5,278 tweets). To the\nbest of our knowledge, METS-CoV is the first dataset to collect medical\nentities and corresponding sentiments of COVID-19-related tweets. We benchmark\nthe performance of classical machine learning models and state-of-the-art deep\nlearning models on NER and TSA tasks with extensive experiments. Results show\nthat the dataset has vast room for improvement for both NER and TSA tasks.\nMETS-CoV is an important resource for developing better medical social media\ntools and facilitating computational social science research, especially in\nepidemiology. Our data, annotation guidelines, benchmark models, and source\ncode are publicly available (https://github.com/YLab-Open/METS-CoV) to ensure\nreproducibility.",
"title": "METS-CoV: A Dataset of Medical Entity and Targeted Sentiment on COVID-19 Related Tweets",
"url": "http://arxiv.org/abs/2209.13773v1"
} | null | null | new_dataset | admin | null | false | null | 401077d3-09e9-4b72-914e-7e0d65b3e45d | null | Validated | {
"text_length": 1979
} | 0new_dataset
|
TITLE: A Benchmarking Dataset with 2440 Organic Molecules for Volume Distribution at Steady State
ABSTRACT: Background: The volume of distribution at steady state (VDss) is a
fundamental pharmacokinetics (PK) property of drugs, which measures how
effectively a drug molecule is distributed throughout the body. Along with the
clearance (CL), it determines the half-life and, therefore, the drug dosing
interval. However, the molecular data size limits the generalizability of the
reported machine learning models. Objective: This study aims to provide a clean
and comprehensive dataset for human VDss as the benchmarking data source,
fostering and benefiting future predictive studies. Moreover, several
predictive models were also built with machine learning regression algorithms.
Methods: The dataset was curated from 13 publicly accessible data sources and
the DrugBank database entirely from intravenous drug administration and then
underwent extensive data cleaning. The molecular descriptors were calculated
with Mordred, and feature selection was conducted for constructing predictive
models. Five machine learning methods were used to build regression models,
grid search was used to optimize hyperparameters, and ten-fold cross-validation
was used to evaluate the model. Results: An enriched dataset of VDss
(https://github.com/da-wen-er/VDss) was constructed with 2440 molecules. Among
the prediction models, the LightGBM model was the most stable and had the best
internal prediction ability with Q2 = 0.837, R2=0.814 and for the other four
models, Q2 was higher than 0.79. Conclusions: To the best of our knowledge,
this is the largest dataset for VDss, which can be used as the benchmark for
computational studies of VDss. Moreover, the regression models reported within
this study can be of use for pharmacokinetic related studies. | {
"abstract": "Background: The volume of distribution at steady state (VDss) is a\nfundamental pharmacokinetics (PK) property of drugs, which measures how\neffectively a drug molecule is distributed throughout the body. Along with the\nclearance (CL), it determines the half-life and, therefore, the drug dosing\ninterval. However, the molecular data size limits the generalizability of the\nreported machine learning models. Objective: This study aims to provide a clean\nand comprehensive dataset for human VDss as the benchmarking data source,\nfostering and benefiting future predictive studies. Moreover, several\npredictive models were also built with machine learning regression algorithms.\nMethods: The dataset was curated from 13 publicly accessible data sources and\nthe DrugBank database entirely from intravenous drug administration and then\nunderwent extensive data cleaning. The molecular descriptors were calculated\nwith Mordred, and feature selection was conducted for constructing predictive\nmodels. Five machine learning methods were used to build regression models,\ngrid search was used to optimize hyperparameters, and ten-fold cross-validation\nwas used to evaluate the model. Results: An enriched dataset of VDss\n(https://github.com/da-wen-er/VDss) was constructed with 2440 molecules. Among\nthe prediction models, the LightGBM model was the most stable and had the best\ninternal prediction ability with Q2 = 0.837, R2=0.814 and for the other four\nmodels, Q2 was higher than 0.79. Conclusions: To the best of our knowledge,\nthis is the largest dataset for VDss, which can be used as the benchmark for\ncomputational studies of VDss. Moreover, the regression models reported within\nthis study can be of use for pharmacokinetic related studies.",
"title": "A Benchmarking Dataset with 2440 Organic Molecules for Volume Distribution at Steady State",
"url": "http://arxiv.org/abs/2211.05661v1"
} | null | null | new_dataset | admin | null | false | null | 4be8cf84-9ee8-452c-8270-486868b8c99f | null | Validated | {
"text_length": 1863
} | 0new_dataset
|
TITLE: SyntheticFur dataset for neural rendering
ABSTRACT: We introduce a new dataset called SyntheticFur built specifically for machine
learning training. The dataset consists of ray traced synthetic fur renders
with corresponding rasterized input buffers and simulation data files. We
procedurally generated approximately 140,000 images and 15 simulations with
Houdini. The images consist of fur groomed with different skin primitives and
move with various motions in a predefined set of lighting environments. We also
demonstrated how the dataset could be used with neural rendering to
significantly improve fur graphics using inexpensive input buffers by training
a conditional generative adversarial network with perceptual loss. We hope the
availability of such high fidelity fur renders will encourage new advances with
neural rendering for a variety of applications. | {
"abstract": "We introduce a new dataset called SyntheticFur built specifically for machine\nlearning training. The dataset consists of ray traced synthetic fur renders\nwith corresponding rasterized input buffers and simulation data files. We\nprocedurally generated approximately 140,000 images and 15 simulations with\nHoudini. The images consist of fur groomed with different skin primitives and\nmove with various motions in a predefined set of lighting environments. We also\ndemonstrated how the dataset could be used with neural rendering to\nsignificantly improve fur graphics using inexpensive input buffers by training\na conditional generative adversarial network with perceptual loss. We hope the\navailability of such high fidelity fur renders will encourage new advances with\nneural rendering for a variety of applications.",
"title": "SyntheticFur dataset for neural rendering",
"url": "http://arxiv.org/abs/2105.06409v1"
} | null | null | new_dataset | admin | null | false | null | 1eb85054-81ea-44c4-ad59-bfe6c5ac9d31 | null | Validated | {
"text_length": 891
} | 0new_dataset
|
TITLE: Seeing the Unseen: Errors and Bias in Visual Datasets
ABSTRACT: From face recognition in smartphones to automatic routing on self-driving
cars, machine vision algorithms lie in the core of these features. These
systems solve image based tasks by identifying and understanding objects,
subsequently making decisions from these information. However, errors in
datasets are usually induced or even magnified in algorithms, at times
resulting in issues such as recognising black people as gorillas and
misrepresenting ethnicities in search results. This paper tracks the errors in
datasets and their impacts, revealing that a flawed dataset could be a result
of limited categories, incomprehensive sourcing and poor classification. | {
"abstract": "From face recognition in smartphones to automatic routing on self-driving\ncars, machine vision algorithms lie in the core of these features. These\nsystems solve image based tasks by identifying and understanding objects,\nsubsequently making decisions from these information. However, errors in\ndatasets are usually induced or even magnified in algorithms, at times\nresulting in issues such as recognising black people as gorillas and\nmisrepresenting ethnicities in search results. This paper tracks the errors in\ndatasets and their impacts, revealing that a flawed dataset could be a result\nof limited categories, incomprehensive sourcing and poor classification.",
"title": "Seeing the Unseen: Errors and Bias in Visual Datasets",
"url": "http://arxiv.org/abs/2211.01847v1"
} | null | null | no_new_dataset | admin | null | false | null | 122d9848-2c69-4636-994b-dc7dcf9f7f5f | null | Validated | {
"text_length": 751
} | 1no_new_dataset
|
TITLE: SC2EGSet: StarCraft II Esport Replay and Game-state Dataset
ABSTRACT: As a relatively new form of sport, esports offers unparalleled data
availability. Despite the vast amounts of data that are generated by game
engines, it can be challenging to extract them and verify their integrity for
the purposes of practical and scientific use.
Our work aims to open esports to a broader scientific community by supplying
raw and pre-processed files from StarCraft II esports tournaments. These files
can be used in statistical and machine learning modeling tasks and related to
various laboratory-based measurements (e.g., behavioral tests, brain imaging).
We have gathered publicly available game-engine generated "replays" of
tournament matches and performed data extraction and cleanup using a low-level
application programming interface (API) parser library.
Additionally, we open-sourced and published all the custom tools that were
developed in the process of creating our dataset. These tools include PyTorch
and PyTorch Lightning API abstractions to load and model the data.
Our dataset contains replays from major and premiere StarCraft II tournaments
since 2016. To prepare the dataset, we processed 55 tournament "replaypacks"
that contained 17930 files with game-state information. Based on initial
investigation of available StarCraft II datasets, we observed that our dataset
is the largest publicly available source of StarCraft II esports data upon its
publication.
Analysis of the extracted data holds promise for further Artificial
Intelligence (AI), Machine Learning (ML), psychological, Human-Computer
Interaction (HCI), and sports-related studies in a variety of supervised and
self-supervised tasks. | {
"abstract": "As a relatively new form of sport, esports offers unparalleled data\navailability. Despite the vast amounts of data that are generated by game\nengines, it can be challenging to extract them and verify their integrity for\nthe purposes of practical and scientific use.\n Our work aims to open esports to a broader scientific community by supplying\nraw and pre-processed files from StarCraft II esports tournaments. These files\ncan be used in statistical and machine learning modeling tasks and related to\nvarious laboratory-based measurements (e.g., behavioral tests, brain imaging).\nWe have gathered publicly available game-engine generated \"replays\" of\ntournament matches and performed data extraction and cleanup using a low-level\napplication programming interface (API) parser library.\n Additionally, we open-sourced and published all the custom tools that were\ndeveloped in the process of creating our dataset. These tools include PyTorch\nand PyTorch Lightning API abstractions to load and model the data.\n Our dataset contains replays from major and premiere StarCraft II tournaments\nsince 2016. To prepare the dataset, we processed 55 tournament \"replaypacks\"\nthat contained 17930 files with game-state information. Based on initial\ninvestigation of available StarCraft II datasets, we observed that our dataset\nis the largest publicly available source of StarCraft II esports data upon its\npublication.\n Analysis of the extracted data holds promise for further Artificial\nIntelligence (AI), Machine Learning (ML), psychological, Human-Computer\nInteraction (HCI), and sports-related studies in a variety of supervised and\nself-supervised tasks.",
"title": "SC2EGSet: StarCraft II Esport Replay and Game-state Dataset",
"url": "http://arxiv.org/abs/2207.03428v2"
} | null | null | new_dataset | admin | null | false | null | d19a263f-324e-451a-8241-d4bbde0e9f3d | null | Validated | {
"text_length": 1745
} | 0new_dataset
|
TITLE: Evaluating a Synthetic Image Dataset Generated with Stable Diffusion
ABSTRACT: We generate synthetic images with the "Stable Diffusion" image generation
model using the Wordnet taxonomy and the definitions of concepts it contains.
This synthetic image database can be used as training data for data
augmentation in machine learning applications, and it is used to investigate
the capabilities of the Stable Diffusion model.
Analyses show that Stable Diffusion can produce correct images for a large
number of concepts, but also a large variety of different representations. The
results show differences depending on the test concepts considered and problems
with very specific concepts. These evaluations were performed using a vision
transformer model for image classification. | {
"abstract": "We generate synthetic images with the \"Stable Diffusion\" image generation\nmodel using the Wordnet taxonomy and the definitions of concepts it contains.\nThis synthetic image database can be used as training data for data\naugmentation in machine learning applications, and it is used to investigate\nthe capabilities of the Stable Diffusion model.\n Analyses show that Stable Diffusion can produce correct images for a large\nnumber of concepts, but also a large variety of different representations. The\nresults show differences depending on the test concepts considered and problems\nwith very specific concepts. These evaluations were performed using a vision\ntransformer model for image classification.",
"title": "Evaluating a Synthetic Image Dataset Generated with Stable Diffusion",
"url": "http://arxiv.org/abs/2211.01777v2"
} | null | null | new_dataset | admin | null | false | null | d51559a2-cac7-49a1-b6e1-bf7fae787356 | null | Validated | {
"text_length": 804
} | 0new_dataset
|
TITLE: ReforesTree: A Dataset for Estimating Tropical Forest Carbon Stock with Deep Learning and Aerial Imagery
ABSTRACT: Forest biomass is a key influence for future climate, and the world urgently
needs highly scalable financing schemes, such as carbon offsetting
certifications, to protect and restore forests. Current manual forest carbon
stock inventory methods of measuring single trees by hand are time, labour, and
cost-intensive and have been shown to be subjective. They can lead to
substantial overestimation of the carbon stock and ultimately distrust in
forest financing. The potential for impact and scale of leveraging advancements
in machine learning and remote sensing technologies is promising but needs to
be of high quality in order to replace the current forest stock protocols for
certifications.
In this paper, we present ReforesTree, a benchmark dataset of forest carbon
stock in six agro-forestry carbon offsetting sites in Ecuador. Furthermore, we
show that a deep learning-based end-to-end model using individual tree
detection from low cost RGB-only drone imagery is accurately estimating forest
carbon stock within official carbon offsetting certification standards.
Additionally, our baseline CNN model outperforms state-of-the-art
satellite-based forest biomass and carbon stock estimates for this type of
small-scale, tropical agro-forestry sites. We present this dataset to encourage
machine learning research in this area to increase accountability and
transparency of monitoring, verification and reporting (MVR) in carbon
offsetting projects, as well as scaling global reforestation financing through
accurate remote sensing. | {
"abstract": "Forest biomass is a key influence for future climate, and the world urgently\nneeds highly scalable financing schemes, such as carbon offsetting\ncertifications, to protect and restore forests. Current manual forest carbon\nstock inventory methods of measuring single trees by hand are time, labour, and\ncost-intensive and have been shown to be subjective. They can lead to\nsubstantial overestimation of the carbon stock and ultimately distrust in\nforest financing. The potential for impact and scale of leveraging advancements\nin machine learning and remote sensing technologies is promising but needs to\nbe of high quality in order to replace the current forest stock protocols for\ncertifications.\n In this paper, we present ReforesTree, a benchmark dataset of forest carbon\nstock in six agro-forestry carbon offsetting sites in Ecuador. Furthermore, we\nshow that a deep learning-based end-to-end model using individual tree\ndetection from low cost RGB-only drone imagery is accurately estimating forest\ncarbon stock within official carbon offsetting certification standards.\nAdditionally, our baseline CNN model outperforms state-of-the-art\nsatellite-based forest biomass and carbon stock estimates for this type of\nsmall-scale, tropical agro-forestry sites. We present this dataset to encourage\nmachine learning research in this area to increase accountability and\ntransparency of monitoring, verification and reporting (MVR) in carbon\noffsetting projects, as well as scaling global reforestation financing through\naccurate remote sensing.",
"title": "ReforesTree: A Dataset for Estimating Tropical Forest Carbon Stock with Deep Learning and Aerial Imagery",
"url": "http://arxiv.org/abs/2201.11192v1"
} | null | null | new_dataset | admin | null | false | null | 55d43471-737a-40d3-b966-0060a4b01cb3 | null | Validated | {
"text_length": 1680
} | 0new_dataset
|
TITLE: Space, Time, and Interaction: A Taxonomy of Corner Cases in Trajectory Datasets for Automated Driving
ABSTRACT: Trajectory data analysis is an essential component for highly automated
driving. Complex models developed with these data predict other road users'
movement and behavior patterns. Based on these predictions - and additional
contextual information such as the course of the road, (traffic) rules, and
interaction with other road users - the highly automated vehicle (HAV) must be
able to reliably and safely perform the task assigned to it, e.g., moving from
point A to B. Ideally, the HAV moves safely through its environment, just as we
would expect a human driver to do. However, if unusual trajectories occur,
so-called trajectory corner cases, a human driver can usually cope well, but an
HAV can quickly get into trouble. In the definition of trajectory corner cases,
which we provide in this work, we will consider the relevance of unusual
trajectories with respect to the task at hand. Based on this, we will also
present a taxonomy of different trajectory corner cases. The categorization of
corner cases into the taxonomy will be shown with examples and is done by cause
and required data sources. To illustrate the complexity between the machine
learning (ML) model and the corner case cause, we present a general processing
chain underlying the taxonomy. | {
"abstract": "Trajectory data analysis is an essential component for highly automated\ndriving. Complex models developed with these data predict other road users'\nmovement and behavior patterns. Based on these predictions - and additional\ncontextual information such as the course of the road, (traffic) rules, and\ninteraction with other road users - the highly automated vehicle (HAV) must be\nable to reliably and safely perform the task assigned to it, e.g., moving from\npoint A to B. Ideally, the HAV moves safely through its environment, just as we\nwould expect a human driver to do. However, if unusual trajectories occur,\nso-called trajectory corner cases, a human driver can usually cope well, but an\nHAV can quickly get into trouble. In the definition of trajectory corner cases,\nwhich we provide in this work, we will consider the relevance of unusual\ntrajectories with respect to the task at hand. Based on this, we will also\npresent a taxonomy of different trajectory corner cases. The categorization of\ncorner cases into the taxonomy will be shown with examples and is done by cause\nand required data sources. To illustrate the complexity between the machine\nlearning (ML) model and the corner case cause, we present a general processing\nchain underlying the taxonomy.",
"title": "Space, Time, and Interaction: A Taxonomy of Corner Cases in Trajectory Datasets for Automated Driving",
"url": "http://arxiv.org/abs/2210.08885v1"
} | null | null | no_new_dataset | admin | null | false | null | 62c45eb5-fb7e-4161-9927-0f3104cea3d9 | null | Validated | {
"text_length": 1401
} | 1no_new_dataset
|
TITLE: Performance of different machine learning methods on activity recognition and pose estimation datasets
ABSTRACT: With advancements in computer vision taking place day by day, recently a lot
of light is being shed on activity recognition. With the range for real-world
applications utilizing this field of study increasing across a multitude of
industries such as security and healthcare, it becomes crucial for businesses
to distinguish which machine learning methods perform better than others in the
area. This paper strives to aid in this predicament i.e. building upon previous
related work, it employs both classical and ensemble approaches on rich pose
estimation (OpenPose) and HAR datasets. Making use of appropriate metrics to
evaluate the performance for each model, the results show that overall, random
forest yields the highest accuracy in classifying ADLs. Relatively all the
models have excellent performance across both datasets, except for logistic
regression and AdaBoost perform poorly in the HAR one. With the limitations of
this paper also discussed in the end, the scope for further research is vast,
which can use this paper as a base in aims of producing better results. | {
"abstract": "With advancements in computer vision taking place day by day, recently a lot\nof light is being shed on activity recognition. With the range for real-world\napplications utilizing this field of study increasing across a multitude of\nindustries such as security and healthcare, it becomes crucial for businesses\nto distinguish which machine learning methods perform better than others in the\narea. This paper strives to aid in this predicament i.e. building upon previous\nrelated work, it employs both classical and ensemble approaches on rich pose\nestimation (OpenPose) and HAR datasets. Making use of appropriate metrics to\nevaluate the performance for each model, the results show that overall, random\nforest yields the highest accuracy in classifying ADLs. Relatively all the\nmodels have excellent performance across both datasets, except for logistic\nregression and AdaBoost perform poorly in the HAR one. With the limitations of\nthis paper also discussed in the end, the scope for further research is vast,\nwhich can use this paper as a base in aims of producing better results.",
"title": "Performance of different machine learning methods on activity recognition and pose estimation datasets",
"url": "http://arxiv.org/abs/2210.10247v1"
} | null | null | no_new_dataset | admin | null | false | null | 19b31832-6e44-40b7-b58d-037fb94a87ca | null | Validated | {
"text_length": 1218
} | 1no_new_dataset
|
TITLE: Model Evaluation in Medical Datasets Over Time
ABSTRACT: Machine learning models deployed in healthcare systems face data drawn from
continually evolving environments. However, researchers proposing such models
typically evaluate them in a time-agnostic manner, with train and test splits
sampling patients throughout the entire study period. We introduce the
Evaluation on Medical Datasets Over Time (EMDOT) framework and Python package,
which evaluates the performance of a model class over time. Across five medical
datasets and a variety of models, we compare two training strategies: (1) using
all historical data, and (2) using a window of the most recent data. We note
changes in performance over time, and identify possible explanations for these
shocks. | {
"abstract": "Machine learning models deployed in healthcare systems face data drawn from\ncontinually evolving environments. However, researchers proposing such models\ntypically evaluate them in a time-agnostic manner, with train and test splits\nsampling patients throughout the entire study period. We introduce the\nEvaluation on Medical Datasets Over Time (EMDOT) framework and Python package,\nwhich evaluates the performance of a model class over time. Across five medical\ndatasets and a variety of models, we compare two training strategies: (1) using\nall historical data, and (2) using a window of the most recent data. We note\nchanges in performance over time, and identify possible explanations for these\nshocks.",
"title": "Model Evaluation in Medical Datasets Over Time",
"url": "http://arxiv.org/abs/2211.07165v1"
} | null | null | no_new_dataset | admin | null | false | null | 14c4fda3-5210-42ac-b939-bbe5e881a6bc | null | Validated | {
"text_length": 786
} | 1no_new_dataset
|
TITLE: Introducing various Semantic Models for Amharic: Experimentation and Evaluation with multiple Tasks and Datasets
ABSTRACT: The availability of different pre-trained semantic models enabled the quick
development of machine learning components for downstream applications. Despite
the availability of abundant text data for low resource languages, only a few
semantic models are publicly available. Publicly available pre-trained models
are usually built as a multilingual version of semantic models that can not fit
well for each language due to context variations. In this work, we introduce
different semantic models for Amharic. After we experiment with the existing
pre-trained semantic models, we trained and fine-tuned nine new different
models using a monolingual text corpus. The models are build using word2Vec
embeddings, distributional thesaurus (DT), contextual embeddings, and DT
embeddings obtained via network embedding algorithms. Moreover, we employ these
models for different NLP tasks and investigate their impact. We find that newly
trained models perform better than pre-trained multilingual models.
Furthermore, models based on contextual embeddings from RoBERTA perform better
than the word2Vec models. | {
"abstract": "The availability of different pre-trained semantic models enabled the quick\ndevelopment of machine learning components for downstream applications. Despite\nthe availability of abundant text data for low resource languages, only a few\nsemantic models are publicly available. Publicly available pre-trained models\nare usually built as a multilingual version of semantic models that can not fit\nwell for each language due to context variations. In this work, we introduce\ndifferent semantic models for Amharic. After we experiment with the existing\npre-trained semantic models, we trained and fine-tuned nine new different\nmodels using a monolingual text corpus. The models are build using word2Vec\nembeddings, distributional thesaurus (DT), contextual embeddings, and DT\nembeddings obtained via network embedding algorithms. Moreover, we employ these\nmodels for different NLP tasks and investigate their impact. We find that newly\ntrained models perform better than pre-trained multilingual models.\nFurthermore, models based on contextual embeddings from RoBERTA perform better\nthan the word2Vec models.",
"title": "Introducing various Semantic Models for Amharic: Experimentation and Evaluation with multiple Tasks and Datasets",
"url": "http://arxiv.org/abs/2011.01154v2"
} | null | null | no_new_dataset | admin | null | false | null | 23f200e3-8943-4963-b563-044769105c27 | null | Validated | {
"text_length": 1248
} | 1no_new_dataset
|
TITLE: QM7-X: A comprehensive dataset of quantum-mechanical properties spanning the chemical space of small organic molecules
ABSTRACT: We introduce QM7-X, a comprehensive dataset of 42 physicochemical properties
for $\approx$ 4.2 M equilibrium and non-equilibrium structures of small organic
molecules with up to seven non-hydrogen (C, N, O, S, Cl) atoms. To span this
fundamentally important region of chemical compound space (CCS), QM7-X includes
an exhaustive sampling of (meta-)stable equilibrium structures - comprised of
constitutional/structural isomers and stereoisomers, e.g., enantiomers and
diastereomers (including cis-/trans- and conformational isomers) - as well as
100 non-equilibrium structural variations thereof to reach a total of $\approx$
4.2 M molecular structures. Computed at the tightly converged
quantum-mechanical PBE0+MBD level of theory, QM7-X contains global (molecular)
and local (atom-in-a-molecule) properties ranging from ground state quantities
(such as atomization energies and dipole moments) to response quantities (such
as polarizability tensors and dispersion coefficients). By providing a
systematic, extensive, and tightly-converged dataset of quantum-mechanically
computed physicochemical properties, we expect that QM7-X will play a critical
role in the development of next-generation machine-learning based models for
exploring greater swaths of CCS and performing in silico design of molecules
with targeted properties. | {
"abstract": "We introduce QM7-X, a comprehensive dataset of 42 physicochemical properties\nfor $\\approx$ 4.2 M equilibrium and non-equilibrium structures of small organic\nmolecules with up to seven non-hydrogen (C, N, O, S, Cl) atoms. To span this\nfundamentally important region of chemical compound space (CCS), QM7-X includes\nan exhaustive sampling of (meta-)stable equilibrium structures - comprised of\nconstitutional/structural isomers and stereoisomers, e.g., enantiomers and\ndiastereomers (including cis-/trans- and conformational isomers) - as well as\n100 non-equilibrium structural variations thereof to reach a total of $\\approx$\n4.2 M molecular structures. Computed at the tightly converged\nquantum-mechanical PBE0+MBD level of theory, QM7-X contains global (molecular)\nand local (atom-in-a-molecule) properties ranging from ground state quantities\n(such as atomization energies and dipole moments) to response quantities (such\nas polarizability tensors and dispersion coefficients). By providing a\nsystematic, extensive, and tightly-converged dataset of quantum-mechanically\ncomputed physicochemical properties, we expect that QM7-X will play a critical\nrole in the development of next-generation machine-learning based models for\nexploring greater swaths of CCS and performing in silico design of molecules\nwith targeted properties.",
"title": "QM7-X: A comprehensive dataset of quantum-mechanical properties spanning the chemical space of small organic molecules",
"url": "http://arxiv.org/abs/2006.15139v1"
} | null | null | new_dataset | admin | null | false | null | 18cb5362-0983-47bd-b347-c514411dc09e | null | Validated | {
"text_length": 1483
} | 0new_dataset
|
TITLE: HealthFC: A Dataset of Health Claims for Evidence-Based Medical Fact-Checking
ABSTRACT: Seeking health-related advice on the internet has become a common practice in
the digital era. Determining the trustworthiness of medical claims found online
and finding appropriate evidence for this information is increasingly
challenging. Fact-checking has emerged as an approach to assess the veracity of
factual claims using evidence from credible knowledge sources. To help advance
the automation of this task, in this paper, we introduce a novel dataset of 750
health-related claims, labeled for veracity by medical experts and backed with
evidence from appropriate clinical studies. We provide an analysis of the
dataset, highlighting its characteristics and challenges. The dataset can be
used for Machine Learning tasks related to automated fact-checking such as
evidence retrieval, veracity prediction, and explanation generation. For this
purpose, we provide baseline models based on different approaches, examine
their performance, and discuss the findings. | {
"abstract": "Seeking health-related advice on the internet has become a common practice in\nthe digital era. Determining the trustworthiness of medical claims found online\nand finding appropriate evidence for this information is increasingly\nchallenging. Fact-checking has emerged as an approach to assess the veracity of\nfactual claims using evidence from credible knowledge sources. To help advance\nthe automation of this task, in this paper, we introduce a novel dataset of 750\nhealth-related claims, labeled for veracity by medical experts and backed with\nevidence from appropriate clinical studies. We provide an analysis of the\ndataset, highlighting its characteristics and challenges. The dataset can be\nused for Machine Learning tasks related to automated fact-checking such as\nevidence retrieval, veracity prediction, and explanation generation. For this\npurpose, we provide baseline models based on different approaches, examine\ntheir performance, and discuss the findings.",
"title": "HealthFC: A Dataset of Health Claims for Evidence-Based Medical Fact-Checking",
"url": "http://arxiv.org/abs/2309.08503v1"
} | null | null | new_dataset | admin | null | false | null | 309ac8b2-3681-43f5-b36c-348a03e320ab | null | Validated | {
"text_length": 1081
} | 0new_dataset
|
TITLE: Sentiment Analysis of Persian Language: Review of Algorithms, Approaches and Datasets
ABSTRACT: Sentiment analysis aims to extract people's emotions and opinion from their
comments on the web. It widely used in businesses to detect sentiment in social
data, gauge brand reputation, and understand customers. Most of articles in
this area have concentrated on the English language whereas there are limited
resources for Persian language. In this review paper, recent published articles
between 2018 and 2022 in sentiment analysis in Persian Language have been
collected and their methods, approach and dataset will be explained and
analyzed. Almost all the methods used to solve sentiment analysis are machine
learning and deep learning. The purpose of this paper is to examine 40
different approach sentiment analysis in the Persian Language, analysis
datasets along with the accuracy of the algorithms applied to them and also
review strengths and weaknesses of each. Among all the methods, transformers
such as BERT and RNN Neural Networks such as LSTM and Bi-LSTM have achieved
higher accuracy in the sentiment analysis. In addition to the methods and
approaches, the datasets reviewed are listed between 2018 and 2022 and
information about each dataset and its details are provided. | {
"abstract": "Sentiment analysis aims to extract people's emotions and opinion from their\ncomments on the web. It widely used in businesses to detect sentiment in social\ndata, gauge brand reputation, and understand customers. Most of articles in\nthis area have concentrated on the English language whereas there are limited\nresources for Persian language. In this review paper, recent published articles\nbetween 2018 and 2022 in sentiment analysis in Persian Language have been\ncollected and their methods, approach and dataset will be explained and\nanalyzed. Almost all the methods used to solve sentiment analysis are machine\nlearning and deep learning. The purpose of this paper is to examine 40\ndifferent approach sentiment analysis in the Persian Language, analysis\ndatasets along with the accuracy of the algorithms applied to them and also\nreview strengths and weaknesses of each. Among all the methods, transformers\nsuch as BERT and RNN Neural Networks such as LSTM and Bi-LSTM have achieved\nhigher accuracy in the sentiment analysis. In addition to the methods and\napproaches, the datasets reviewed are listed between 2018 and 2022 and\ninformation about each dataset and its details are provided.",
"title": "Sentiment Analysis of Persian Language: Review of Algorithms, Approaches and Datasets",
"url": "http://arxiv.org/abs/2212.06041v1"
} | null | null | no_new_dataset | admin | null | false | null | 28a03d50-7c85-4a8b-aea1-a81c38f5012b | null | Validated | {
"text_length": 1311
} | 1no_new_dataset
|
TITLE: Crime Prediction using Machine Learning with a Novel Crime Dataset
ABSTRACT: Crime is an unlawful act that carries legal repercussions. Bangladesh has a
high crime rate due to poverty, population growth, and many other
socio-economic issues. For law enforcement agencies, understanding crime
patterns is essential for preventing future criminal activity. For this
purpose, these agencies need structured crime database. This paper introduces a
novel crime dataset that contains temporal, geographic, weather, and
demographic data about 6574 crime incidents of Bangladesh. We manually gather
crime news articles of a seven year time span from a daily newspaper archive.
We extract basic features from these raw text. Using these basic features, we
then consult standard service-providers of geo-location and weather data in
order to garner these information related to the collected crime incidents.
Furthermore, we collect demographic information from Bangladesh National Census
data. All these information are combined that results in a standard machine
learning dataset. Together, 36 features are engineered for the crime prediction
task. Five supervised machine learning classification algorithms are then
evaluated on this newly built dataset and satisfactory results are achieved. We
also conduct exploratory analysis on various aspects the dataset. This dataset
is expected to serve as the foundation for crime incidence prediction systems
for Bangladesh and other countries. The findings of this study will help law
enforcement agencies to forecast and contain crime as well as to ensure optimal
resource allocation for crime patrol and prevention. | {
"abstract": "Crime is an unlawful act that carries legal repercussions. Bangladesh has a\nhigh crime rate due to poverty, population growth, and many other\nsocio-economic issues. For law enforcement agencies, understanding crime\npatterns is essential for preventing future criminal activity. For this\npurpose, these agencies need structured crime database. This paper introduces a\nnovel crime dataset that contains temporal, geographic, weather, and\ndemographic data about 6574 crime incidents of Bangladesh. We manually gather\ncrime news articles of a seven year time span from a daily newspaper archive.\nWe extract basic features from these raw text. Using these basic features, we\nthen consult standard service-providers of geo-location and weather data in\norder to garner these information related to the collected crime incidents.\nFurthermore, we collect demographic information from Bangladesh National Census\ndata. All these information are combined that results in a standard machine\nlearning dataset. Together, 36 features are engineered for the crime prediction\ntask. Five supervised machine learning classification algorithms are then\nevaluated on this newly built dataset and satisfactory results are achieved. We\nalso conduct exploratory analysis on various aspects the dataset. This dataset\nis expected to serve as the foundation for crime incidence prediction systems\nfor Bangladesh and other countries. The findings of this study will help law\nenforcement agencies to forecast and contain crime as well as to ensure optimal\nresource allocation for crime patrol and prevention.",
"title": "Crime Prediction using Machine Learning with a Novel Crime Dataset",
"url": "http://arxiv.org/abs/2211.01551v1"
} | null | null | new_dataset | admin | null | false | null | 10f466ee-7a83-42d0-911f-27a639abf0ee | null | Validated | {
"text_length": 1679
} | 0new_dataset
|
TITLE: A universal synthetic dataset for machine learning on spectroscopic data
ABSTRACT: To assist in the development of machine learning methods for automated
classification of spectroscopic data, we have generated a universal synthetic
dataset that can be used for model validation. This dataset contains artificial
spectra designed to represent experimental measurements from techniques
including X-ray diffraction, nuclear magnetic resonance, and Raman
spectroscopy. The dataset generation process features customizable parameters,
such as scan length and peak count, which can be adjusted to fit the problem at
hand. As an initial benchmark, we simulated a dataset containing 35,000 spectra
based on 500 unique classes. To automate the classification of this data, eight
different machine learning architectures were evaluated. From the results, we
shed light on which factors are most critical to achieve optimal performance
for the classification task. The scripts used to generate synthetic spectra, as
well as our benchmark dataset and evaluation routines, are made publicly
available to aid in the development of improved machine learning models for
spectroscopic analysis. | {
"abstract": "To assist in the development of machine learning methods for automated\nclassification of spectroscopic data, we have generated a universal synthetic\ndataset that can be used for model validation. This dataset contains artificial\nspectra designed to represent experimental measurements from techniques\nincluding X-ray diffraction, nuclear magnetic resonance, and Raman\nspectroscopy. The dataset generation process features customizable parameters,\nsuch as scan length and peak count, which can be adjusted to fit the problem at\nhand. As an initial benchmark, we simulated a dataset containing 35,000 spectra\nbased on 500 unique classes. To automate the classification of this data, eight\ndifferent machine learning architectures were evaluated. From the results, we\nshed light on which factors are most critical to achieve optimal performance\nfor the classification task. The scripts used to generate synthetic spectra, as\nwell as our benchmark dataset and evaluation routines, are made publicly\navailable to aid in the development of improved machine learning models for\nspectroscopic analysis.",
"title": "A universal synthetic dataset for machine learning on spectroscopic data",
"url": "http://arxiv.org/abs/2206.06031v2"
} | null | null | new_dataset | admin | null | false | null | 3b5aa46c-1600-414e-96b9-4f0a87a2e7ba | null | Validated | {
"text_length": 1201
} | 0new_dataset
|
TITLE: Minimizing the Effect of Noise and Limited Dataset Size in Image Classification Using Depth Estimation as an Auxiliary Task with Deep Multitask Learning
ABSTRACT: Generalizability is the ultimate goal of Machine Learning (ML) image
classifiers, for which noise and limited dataset size are among the major
concerns. We tackle these challenges through utilizing the framework of deep
Multitask Learning (dMTL) and incorporating image depth estimation as an
auxiliary task. On a customized and depth-augmented derivation of the MNIST
dataset, we show a) multitask loss functions are the most effective approach of
implementing dMTL, b) limited dataset size primarily contributes to
classification inaccuracy, and c) depth estimation is mostly impacted by noise.
In order to further validate the results, we manually labeled the NYU Depth V2
dataset for scene classification tasks. As a contribution to the field, we have
made the data in python native format publicly available as an open-source
dataset and provided the scene labels. Our experiments on MNIST and
NYU-Depth-V2 show dMTL improves generalizability of the classifiers when the
dataset is noisy and the number of examples is limited. | {
"abstract": "Generalizability is the ultimate goal of Machine Learning (ML) image\nclassifiers, for which noise and limited dataset size are among the major\nconcerns. We tackle these challenges through utilizing the framework of deep\nMultitask Learning (dMTL) and incorporating image depth estimation as an\nauxiliary task. On a customized and depth-augmented derivation of the MNIST\ndataset, we show a) multitask loss functions are the most effective approach of\nimplementing dMTL, b) limited dataset size primarily contributes to\nclassification inaccuracy, and c) depth estimation is mostly impacted by noise.\nIn order to further validate the results, we manually labeled the NYU Depth V2\ndataset for scene classification tasks. As a contribution to the field, we have\nmade the data in python native format publicly available as an open-source\ndataset and provided the scene labels. Our experiments on MNIST and\nNYU-Depth-V2 show dMTL improves generalizability of the classifiers when the\ndataset is noisy and the number of examples is limited.",
"title": "Minimizing the Effect of Noise and Limited Dataset Size in Image Classification Using Depth Estimation as an Auxiliary Task with Deep Multitask Learning",
"url": "http://arxiv.org/abs/2208.10390v1"
} | null | null | no_new_dataset | admin | null | false | null | a31f3628-2185-4f85-8b93-632125fa0c3d | null | Validated | {
"text_length": 1218
} | 1no_new_dataset
|
TITLE: Video compression dataset and benchmark of learning-based video-quality metrics
ABSTRACT: Video-quality measurement is a critical task in video processing. Nowadays,
many implementations of new encoding standards - such as AV1, VVC, and LCEVC -
use deep-learning-based decoding algorithms with perceptual metrics that serve
as optimization objectives. But investigations of the performance of modern
video- and image-quality metrics commonly employ videos compressed using older
standards, such as AVC. In this paper, we present a new benchmark for
video-quality metrics that evaluates video compression. It is based on a new
dataset consisting of about 2,500 streams encoded using different standards,
including AVC, HEVC, AV1, VP9, and VVC. Subjective scores were collected using
crowdsourced pairwise comparisons. The list of evaluated metrics includes
recent ones based on machine learning and neural networks. The results
demonstrate that new no-reference metrics exhibit a high correlation with
subjective quality and approach the capability of top full-reference metrics. | {
"abstract": "Video-quality measurement is a critical task in video processing. Nowadays,\nmany implementations of new encoding standards - such as AV1, VVC, and LCEVC -\nuse deep-learning-based decoding algorithms with perceptual metrics that serve\nas optimization objectives. But investigations of the performance of modern\nvideo- and image-quality metrics commonly employ videos compressed using older\nstandards, such as AVC. In this paper, we present a new benchmark for\nvideo-quality metrics that evaluates video compression. It is based on a new\ndataset consisting of about 2,500 streams encoded using different standards,\nincluding AVC, HEVC, AV1, VP9, and VVC. Subjective scores were collected using\ncrowdsourced pairwise comparisons. The list of evaluated metrics includes\nrecent ones based on machine learning and neural networks. The results\ndemonstrate that new no-reference metrics exhibit a high correlation with\nsubjective quality and approach the capability of top full-reference metrics.",
"title": "Video compression dataset and benchmark of learning-based video-quality metrics",
"url": "http://arxiv.org/abs/2211.12109v2"
} | null | null | new_dataset | admin | null | false | null | 16bdbcde-ffba-4939-952e-71b06be560eb | null | Validated | {
"text_length": 1102
} | 0new_dataset
|
TITLE: Multi Visual Modality Fall Detection Dataset
ABSTRACT: Falls are one of the leading cause of injury-related deaths among the elderly
worldwide. Effective detection of falls can reduce the risk of complications
and injuries. Fall detection can be performed using wearable devices or ambient
sensors; these methods may struggle with user compliance issues or false
alarms. Video cameras provide a passive alternative; however, regular RGB
cameras are impacted by changing lighting conditions and privacy concerns. From
a machine learning perspective, developing an effective fall detection system
is challenging because of the rarity and variability of falls. Many existing
fall detection datasets lack important real-world considerations, such as
varied lighting, continuous activities of daily living (ADLs), and camera
placement. The lack of these considerations makes it difficult to develop
predictive models that can operate effectively in the real world. To address
these limitations, we introduce a novel multi-modality dataset (MUVIM) that
contains four visual modalities: infra-red, depth, RGB and thermal cameras.
These modalities offer benefits such as obfuscated facial features and improved
performance in low-light conditions. We formulated fall detection as an anomaly
detection problem, in which a customized spatio-temporal convolutional
autoencoder was trained only on ADLs so that a fall would increase the
reconstruction error. Our results showed that infra-red cameras provided the
highest level of performance (AUC ROC=0.94), followed by thermal (AUC
ROC=0.87), depth (AUC ROC=0.86) and RGB (AUC ROC=0.83). This research provides
a unique opportunity to analyze the utility of camera modalities in detecting
falls in a home setting while balancing performance, passiveness, and privacy. | {
"abstract": "Falls are one of the leading cause of injury-related deaths among the elderly\nworldwide. Effective detection of falls can reduce the risk of complications\nand injuries. Fall detection can be performed using wearable devices or ambient\nsensors; these methods may struggle with user compliance issues or false\nalarms. Video cameras provide a passive alternative; however, regular RGB\ncameras are impacted by changing lighting conditions and privacy concerns. From\na machine learning perspective, developing an effective fall detection system\nis challenging because of the rarity and variability of falls. Many existing\nfall detection datasets lack important real-world considerations, such as\nvaried lighting, continuous activities of daily living (ADLs), and camera\nplacement. The lack of these considerations makes it difficult to develop\npredictive models that can operate effectively in the real world. To address\nthese limitations, we introduce a novel multi-modality dataset (MUVIM) that\ncontains four visual modalities: infra-red, depth, RGB and thermal cameras.\nThese modalities offer benefits such as obfuscated facial features and improved\nperformance in low-light conditions. We formulated fall detection as an anomaly\ndetection problem, in which a customized spatio-temporal convolutional\nautoencoder was trained only on ADLs so that a fall would increase the\nreconstruction error. Our results showed that infra-red cameras provided the\nhighest level of performance (AUC ROC=0.94), followed by thermal (AUC\nROC=0.87), depth (AUC ROC=0.86) and RGB (AUC ROC=0.83). This research provides\na unique opportunity to analyze the utility of camera modalities in detecting\nfalls in a home setting while balancing performance, passiveness, and privacy.",
"title": "Multi Visual Modality Fall Detection Dataset",
"url": "http://arxiv.org/abs/2206.12740v1"
} | null | null | new_dataset | admin | null | false | null | 9e11cce3-f6fa-437c-9be7-a30019f60463 | null | Validated | {
"text_length": 1831
} | 0new_dataset
|
TITLE: BD-SHS: A Benchmark Dataset for Learning to Detect Online Bangla Hate Speech in Different Social Contexts
ABSTRACT: Social media platforms and online streaming services have spawned a new breed
of Hate Speech (HS). Due to the massive amount of user-generated content on
these sites, modern machine learning techniques are found to be feasible and
cost-effective to tackle this problem. However, linguistically diverse datasets
covering different social contexts in which offensive language is typically
used are required to train generalizable models. In this paper, we identify the
shortcomings of existing Bangla HS datasets and introduce a large manually
labeled dataset BD-SHS that includes HS in different social contexts. The
labeling criteria were prepared following a hierarchical annotation process,
which is the first of its kind in Bangla HS to the best of our knowledge. The
dataset includes more than 50,200 offensive comments crawled from online social
networking sites and is at least 60% larger than any existing Bangla HS
datasets. We present the benchmark result of our dataset by training different
NLP models resulting in the best one achieving an F1-score of 91.0%. In our
experiments, we found that a word embedding trained exclusively using 1.47
million comments from social media and streaming sites consistently resulted in
better modeling of HS detection in comparison to other pre-trained embeddings.
Our dataset and all accompanying codes is publicly available at
github.com/naurosromim/hate-speech-dataset-for-Bengali-social-media | {
"abstract": "Social media platforms and online streaming services have spawned a new breed\nof Hate Speech (HS). Due to the massive amount of user-generated content on\nthese sites, modern machine learning techniques are found to be feasible and\ncost-effective to tackle this problem. However, linguistically diverse datasets\ncovering different social contexts in which offensive language is typically\nused are required to train generalizable models. In this paper, we identify the\nshortcomings of existing Bangla HS datasets and introduce a large manually\nlabeled dataset BD-SHS that includes HS in different social contexts. The\nlabeling criteria were prepared following a hierarchical annotation process,\nwhich is the first of its kind in Bangla HS to the best of our knowledge. The\ndataset includes more than 50,200 offensive comments crawled from online social\nnetworking sites and is at least 60% larger than any existing Bangla HS\ndatasets. We present the benchmark result of our dataset by training different\nNLP models resulting in the best one achieving an F1-score of 91.0%. In our\nexperiments, we found that a word embedding trained exclusively using 1.47\nmillion comments from social media and streaming sites consistently resulted in\nbetter modeling of HS detection in comparison to other pre-trained embeddings.\nOur dataset and all accompanying codes is publicly available at\ngithub.com/naurosromim/hate-speech-dataset-for-Bengali-social-media",
"title": "BD-SHS: A Benchmark Dataset for Learning to Detect Online Bangla Hate Speech in Different Social Contexts",
"url": "http://arxiv.org/abs/2206.00372v1"
} | null | null | new_dataset | admin | null | false | null | 08be5b95-ee31-4168-9bce-0a703f02d5af | null | Validated | {
"text_length": 1583
} | 0new_dataset
|
TITLE: MultiWOZ-DF -- A Dataflow implementation of the MultiWOZ dataset
ABSTRACT: Semantic Machines (SM) have introduced the use of the dataflow (DF) paradigm
to dialogue modelling, using computational graphs to hierarchically represent
user requests, data, and the dialogue history [Semantic Machines et al. 2020].
Although the main focus of that paper was the SMCalFlow dataset (to date, the
only dataset with "native" DF annotations), they also reported some results of
an experiment using a transformed version of the commonly used MultiWOZ dataset
[Budzianowski et al. 2018] into a DF format. In this paper, we expand the
experiments using DF for the MultiWOZ dataset, exploring some additional
experimental set-ups. The code and instructions to reproduce the experiments
reported here have been released. The contributions of this paper are: 1.) A DF
implementation capable of executing MultiWOZ dialogues; 2.) Several versions of
conversion of MultiWOZ into a DF format are presented; 3.) Experimental results
on state match and translation accuracy. | {
"abstract": "Semantic Machines (SM) have introduced the use of the dataflow (DF) paradigm\nto dialogue modelling, using computational graphs to hierarchically represent\nuser requests, data, and the dialogue history [Semantic Machines et al. 2020].\nAlthough the main focus of that paper was the SMCalFlow dataset (to date, the\nonly dataset with \"native\" DF annotations), they also reported some results of\nan experiment using a transformed version of the commonly used MultiWOZ dataset\n[Budzianowski et al. 2018] into a DF format. In this paper, we expand the\nexperiments using DF for the MultiWOZ dataset, exploring some additional\nexperimental set-ups. The code and instructions to reproduce the experiments\nreported here have been released. The contributions of this paper are: 1.) A DF\nimplementation capable of executing MultiWOZ dialogues; 2.) Several versions of\nconversion of MultiWOZ into a DF format are presented; 3.) Experimental results\non state match and translation accuracy.",
"title": "MultiWOZ-DF -- A Dataflow implementation of the MultiWOZ dataset",
"url": "http://arxiv.org/abs/2211.02303v1"
} | null | null | no_new_dataset | admin | null | false | null | 179b9cff-02d5-4dec-9f40-b2cd7e6d34eb | null | Validated | {
"text_length": 1074
} | 1no_new_dataset
|
TITLE: DSLOB: A Synthetic Limit Order Book Dataset for Benchmarking Forecasting Algorithms under Distributional Shift
ABSTRACT: In electronic trading markets, limit order books (LOBs) provide information
about pending buy/sell orders at various price levels for a given security.
Recently, there has been a growing interest in using LOB data for resolving
downstream machine learning tasks (e.g., forecasting). However, dealing with
out-of-distribution (OOD) LOB data is challenging since distributional shifts
are unlabeled in current publicly available LOB datasets. Therefore, it is
critical to build a synthetic LOB dataset with labeled OOD samples serving as a
testbed for developing models that generalize well to unseen scenarios. In this
work, we utilize a multi-agent market simulator to build a synthetic LOB
dataset, named DSLOB, with and without market stress scenarios, which allows
for the design of controlled distributional shift benchmarking. Using the
proposed synthetic dataset, we provide a holistic analysis on the forecasting
performance of three different state-of-the-art forecasting methods. Our
results reflect the need for increased researcher efforts to develop algorithms
with robustness to distributional shifts in high-frequency time series data. | {
"abstract": "In electronic trading markets, limit order books (LOBs) provide information\nabout pending buy/sell orders at various price levels for a given security.\nRecently, there has been a growing interest in using LOB data for resolving\ndownstream machine learning tasks (e.g., forecasting). However, dealing with\nout-of-distribution (OOD) LOB data is challenging since distributional shifts\nare unlabeled in current publicly available LOB datasets. Therefore, it is\ncritical to build a synthetic LOB dataset with labeled OOD samples serving as a\ntestbed for developing models that generalize well to unseen scenarios. In this\nwork, we utilize a multi-agent market simulator to build a synthetic LOB\ndataset, named DSLOB, with and without market stress scenarios, which allows\nfor the design of controlled distributional shift benchmarking. Using the\nproposed synthetic dataset, we provide a holistic analysis on the forecasting\nperformance of three different state-of-the-art forecasting methods. Our\nresults reflect the need for increased researcher efforts to develop algorithms\nwith robustness to distributional shifts in high-frequency time series data.",
"title": "DSLOB: A Synthetic Limit Order Book Dataset for Benchmarking Forecasting Algorithms under Distributional Shift",
"url": "http://arxiv.org/abs/2211.11513v1"
} | null | null | new_dataset | admin | null | false | null | 78587d52-8637-42a4-8361-723f7f996860 | null | Validated | {
"text_length": 1294
} | 0new_dataset
|
TITLE: Hyperparameter Importance of Quantum Neural Networks Across Small Datasets
ABSTRACT: As restricted quantum computers are slowly becoming a reality, the search for
meaningful first applications intensifies. In this domain, one of the more
investigated approaches is the use of a special type of quantum circuit - a
so-called quantum neural network -- to serve as a basis for a machine learning
model. Roughly speaking, as the name suggests, a quantum neural network can
play a similar role to a neural network. However, specifically for applications
in machine learning contexts, very little is known about suitable circuit
architectures, or model hyperparameters one should use to achieve good learning
performance. In this work, we apply the functional ANOVA framework to quantum
neural networks to analyze which of the hyperparameters were most influential
for their predictive performance. We analyze one of the most typically used
quantum neural network architectures. We then apply this to $7$ open-source
datasets from the OpenML-CC18 classification benchmark whose number of features
is small enough to fit on quantum hardware with less than $20$ qubits. Three
main levels of importance were detected from the ranking of hyperparameters
obtained with functional ANOVA. Our experiment both confirmed expected patterns
and revealed new insights. For instance, setting well the learning rate is
deemed the most critical hyperparameter in terms of marginal contribution on
all datasets, whereas the particular choice of entangling gates used is
considered the least important except on one dataset. This work introduces new
methodologies to study quantum machine learning models and provides new
insights toward quantum model selection. | {
"abstract": "As restricted quantum computers are slowly becoming a reality, the search for\nmeaningful first applications intensifies. In this domain, one of the more\ninvestigated approaches is the use of a special type of quantum circuit - a\nso-called quantum neural network -- to serve as a basis for a machine learning\nmodel. Roughly speaking, as the name suggests, a quantum neural network can\nplay a similar role to a neural network. However, specifically for applications\nin machine learning contexts, very little is known about suitable circuit\narchitectures, or model hyperparameters one should use to achieve good learning\nperformance. In this work, we apply the functional ANOVA framework to quantum\nneural networks to analyze which of the hyperparameters were most influential\nfor their predictive performance. We analyze one of the most typically used\nquantum neural network architectures. We then apply this to $7$ open-source\ndatasets from the OpenML-CC18 classification benchmark whose number of features\nis small enough to fit on quantum hardware with less than $20$ qubits. Three\nmain levels of importance were detected from the ranking of hyperparameters\nobtained with functional ANOVA. Our experiment both confirmed expected patterns\nand revealed new insights. For instance, setting well the learning rate is\ndeemed the most critical hyperparameter in terms of marginal contribution on\nall datasets, whereas the particular choice of entangling gates used is\nconsidered the least important except on one dataset. This work introduces new\nmethodologies to study quantum machine learning models and provides new\ninsights toward quantum model selection.",
"title": "Hyperparameter Importance of Quantum Neural Networks Across Small Datasets",
"url": "http://arxiv.org/abs/2206.09992v1"
} | null | null | no_new_dataset | admin | null | false | null | bfc8fcf3-486b-43bb-97d2-204ccbba07b4 | null | Validated | {
"text_length": 1763
} | 1no_new_dataset
|
TITLE: TweetDIS: A Large Twitter Dataset for Natural Disasters Built using Weak Supervision
ABSTRACT: Social media is often utilized as a lifeline for communication during natural
disasters. Traditionally, natural disaster tweets are filtered from the Twitter
stream using the name of the natural disaster and the filtered tweets are sent
for human annotation. The process of human annotation to create labeled sets
for machine learning models is laborious, time consuming, at times inaccurate,
and more importantly not scalable in terms of size and real-time use. In this
work, we curate a silver standard dataset using weak supervision. In order to
validate its utility, we train machine learning models on the weakly supervised
data to identify three different types of natural disasters i.e earthquakes,
hurricanes and floods. Our results demonstrate that models trained on the
silver standard dataset achieved performance greater than 90% when classifying
a manually curated, gold-standard dataset. To enable reproducible research and
additional downstream utility, we release the silver standard dataset for the
scientific community. | {
"abstract": "Social media is often utilized as a lifeline for communication during natural\ndisasters. Traditionally, natural disaster tweets are filtered from the Twitter\nstream using the name of the natural disaster and the filtered tweets are sent\nfor human annotation. The process of human annotation to create labeled sets\nfor machine learning models is laborious, time consuming, at times inaccurate,\nand more importantly not scalable in terms of size and real-time use. In this\nwork, we curate a silver standard dataset using weak supervision. In order to\nvalidate its utility, we train machine learning models on the weakly supervised\ndata to identify three different types of natural disasters i.e earthquakes,\nhurricanes and floods. Our results demonstrate that models trained on the\nsilver standard dataset achieved performance greater than 90% when classifying\na manually curated, gold-standard dataset. To enable reproducible research and\nadditional downstream utility, we release the silver standard dataset for the\nscientific community.",
"title": "TweetDIS: A Large Twitter Dataset for Natural Disasters Built using Weak Supervision",
"url": "http://arxiv.org/abs/2207.04947v1"
} | null | null | new_dataset | admin | null | false | null | cee0e736-5c41-4882-affa-1e27cafad9b2 | null | Validated | {
"text_length": 1156
} | 0new_dataset
|
TITLE: MANTRA: A Machine Learning reference lightcurve dataset for astronomical transient event recognition
ABSTRACT: We introduce MANTRA, an annotated dataset of 4869 transient and 71207
non-transient object lightcurves built from the Catalina Real Time Transient
Survey. We provide public access to this dataset as a plain text file to
facilitate standardized quantitative comparison of astronomical transient event
recognition algorithms. Some of the classes included in the dataset are:
supernovae, cataclysmic variables, active galactic nuclei, high proper motion
stars, blazars and flares. As an example of the tasks that can be performed on
the dataset we experiment with multiple data pre-processing methods, feature
selection techniques and popular machine learning algorithms (Support Vector
Machines, Random Forests and Neural Networks). We assess quantitative
performance in two classification tasks: binary (transient/non-transient) and
eight-class classification. The best performing algorithm in both tasks is the
Random Forest Classifier. It achieves an F1-score of 96.25% in the binary
classification and 52.79% in the eight-class classification. For the
eight-class classification, non-transients ( 96.83% ) is the class with the
highest F1-score, while the lowest corresponds to high-proper-motion stars (
16.79% ); for supernovae it achieves a value of 54.57% , close to the average
across classes. The next release of MANTRA includes images and benchmarks with
deep learning models. | {
"abstract": "We introduce MANTRA, an annotated dataset of 4869 transient and 71207\nnon-transient object lightcurves built from the Catalina Real Time Transient\nSurvey. We provide public access to this dataset as a plain text file to\nfacilitate standardized quantitative comparison of astronomical transient event\nrecognition algorithms. Some of the classes included in the dataset are:\nsupernovae, cataclysmic variables, active galactic nuclei, high proper motion\nstars, blazars and flares. As an example of the tasks that can be performed on\nthe dataset we experiment with multiple data pre-processing methods, feature\nselection techniques and popular machine learning algorithms (Support Vector\nMachines, Random Forests and Neural Networks). We assess quantitative\nperformance in two classification tasks: binary (transient/non-transient) and\neight-class classification. The best performing algorithm in both tasks is the\nRandom Forest Classifier. It achieves an F1-score of 96.25% in the binary\nclassification and 52.79% in the eight-class classification. For the\neight-class classification, non-transients ( 96.83% ) is the class with the\nhighest F1-score, while the lowest corresponds to high-proper-motion stars (\n16.79% ); for supernovae it achieves a value of 54.57% , close to the average\nacross classes. The next release of MANTRA includes images and benchmarks with\ndeep learning models.",
"title": "MANTRA: A Machine Learning reference lightcurve dataset for astronomical transient event recognition",
"url": "http://arxiv.org/abs/2006.13163v2"
} | null | null | new_dataset | admin | null | false | null | 25fda4a6-8548-4254-8c07-26121b7d7c20 | null | Validated | {
"text_length": 1520
} | 0new_dataset
|
TITLE: The ArtBench Dataset: Benchmarking Generative Models with Artworks
ABSTRACT: We introduce ArtBench-10, the first class-balanced, high-quality, cleanly
annotated, and standardized dataset for benchmarking artwork generation. It
comprises 60,000 images of artwork from 10 distinctive artistic styles, with
5,000 training images and 1,000 testing images per style. ArtBench-10 has
several advantages over previous artwork datasets. Firstly, it is
class-balanced while most previous artwork datasets suffer from the long tail
class distributions. Secondly, the images are of high quality with clean
annotations. Thirdly, ArtBench-10 is created with standardized data collection,
annotation, filtering, and preprocessing procedures. We provide three versions
of the dataset with different resolutions ($32\times32$, $256\times256$, and
original image size), formatted in a way that is easy to be incorporated by
popular machine learning frameworks. We also conduct extensive benchmarking
experiments using representative image synthesis models with ArtBench-10 and
present in-depth analysis. The dataset is available at
https://github.com/liaopeiyuan/artbench under a Fair Use license. | {
"abstract": "We introduce ArtBench-10, the first class-balanced, high-quality, cleanly\nannotated, and standardized dataset for benchmarking artwork generation. It\ncomprises 60,000 images of artwork from 10 distinctive artistic styles, with\n5,000 training images and 1,000 testing images per style. ArtBench-10 has\nseveral advantages over previous artwork datasets. Firstly, it is\nclass-balanced while most previous artwork datasets suffer from the long tail\nclass distributions. Secondly, the images are of high quality with clean\nannotations. Thirdly, ArtBench-10 is created with standardized data collection,\nannotation, filtering, and preprocessing procedures. We provide three versions\nof the dataset with different resolutions ($32\\times32$, $256\\times256$, and\noriginal image size), formatted in a way that is easy to be incorporated by\npopular machine learning frameworks. We also conduct extensive benchmarking\nexperiments using representative image synthesis models with ArtBench-10 and\npresent in-depth analysis. The dataset is available at\nhttps://github.com/liaopeiyuan/artbench under a Fair Use license.",
"title": "The ArtBench Dataset: Benchmarking Generative Models with Artworks",
"url": "http://arxiv.org/abs/2206.11404v1"
} | null | null | new_dataset | admin | null | false | null | 7e646a05-d5a1-426a-8b26-65595dc49b95 | null | Validated | {
"text_length": 1204
} | 0new_dataset
|
TITLE: OpenPack: A Large-scale Dataset for Recognizing Packaging Works in IoT-enabled Logistic Environments
ABSTRACT: Unlike human daily activities, existing publicly available sensor datasets
for work activity recognition in industrial domains are limited by difficulties
in collecting realistic data as close collaboration with industrial sites is
required. This also limits research on and development of AI methods for
industrial applications. To address these challenges and contribute to research
on machine recognition of work activities in industrial domains, in this study,
we introduce a new large-scale dataset for packaging work recognition called
OpenPack. OpenPack contains 53.8 hours of multimodal sensor data, including
keypoints, depth images, acceleration data, and readings from IoT-enabled
devices (e.g., handheld barcode scanners used in work procedures), collected
from 16 distinct subjects with different levels of packaging work experience.
On the basis of this dataset, we propose a neural network model designed to
recognize work activities, which efficiently fuses sensor data and readings
from IoT-enabled devices by processing them within different streams in a
ladder-shaped architecture, and the experiment showed the effectiveness of the
architecture. We believe that OpenPack will contribute to the community of
action/activity recognition with sensors. OpenPack dataset is available at
https://open-pack.github.io/. | {
"abstract": "Unlike human daily activities, existing publicly available sensor datasets\nfor work activity recognition in industrial domains are limited by difficulties\nin collecting realistic data as close collaboration with industrial sites is\nrequired. This also limits research on and development of AI methods for\nindustrial applications. To address these challenges and contribute to research\non machine recognition of work activities in industrial domains, in this study,\nwe introduce a new large-scale dataset for packaging work recognition called\nOpenPack. OpenPack contains 53.8 hours of multimodal sensor data, including\nkeypoints, depth images, acceleration data, and readings from IoT-enabled\ndevices (e.g., handheld barcode scanners used in work procedures), collected\nfrom 16 distinct subjects with different levels of packaging work experience.\nOn the basis of this dataset, we propose a neural network model designed to\nrecognize work activities, which efficiently fuses sensor data and readings\nfrom IoT-enabled devices by processing them within different streams in a\nladder-shaped architecture, and the experiment showed the effectiveness of the\narchitecture. We believe that OpenPack will contribute to the community of\naction/activity recognition with sensors. OpenPack dataset is available at\nhttps://open-pack.github.io/.",
"title": "OpenPack: A Large-scale Dataset for Recognizing Packaging Works in IoT-enabled Logistic Environments",
"url": "http://arxiv.org/abs/2212.11152v1"
} | null | null | new_dataset | admin | null | false | null | 346b23ab-dcef-4472-aa02-30e3ef237894 | null | Validated | {
"text_length": 1466
} | 0new_dataset
|
TITLE: PEOPL: Characterizing Privately Encoded Open Datasets with Public Labels
ABSTRACT: Allowing organizations to share their data for training of machine learning
(ML) models without unintended information leakage is an open problem in
practice. A promising technique for this still-open problem is to train models
on the encoded data. Our approach, called Privately Encoded Open Datasets with
Public Labels (PEOPL), uses a certain class of randomly constructed transforms
to encode sensitive data. Organizations publish their randomly encoded data and
associated raw labels for ML training, where training is done without knowledge
of the encoding realization. We investigate several important aspects of this
problem: We introduce information-theoretic scores for privacy and utility,
which quantify the average performance of an unfaithful user (e.g., adversary)
and a faithful user (e.g., model developer) that have access to the published
encoded data. We then theoretically characterize primitives in building
families of encoding schemes that motivate the use of random deep neural
networks. Empirically, we compare the performance of our randomized encoding
scheme and a linear scheme to a suite of computational attacks, and we also
show that our scheme achieves competitive prediction accuracy to raw-sample
baselines. Moreover, we demonstrate that multiple institutions, using
independent random encoders, can collaborate to train improved ML models. | {
"abstract": "Allowing organizations to share their data for training of machine learning\n(ML) models without unintended information leakage is an open problem in\npractice. A promising technique for this still-open problem is to train models\non the encoded data. Our approach, called Privately Encoded Open Datasets with\nPublic Labels (PEOPL), uses a certain class of randomly constructed transforms\nto encode sensitive data. Organizations publish their randomly encoded data and\nassociated raw labels for ML training, where training is done without knowledge\nof the encoding realization. We investigate several important aspects of this\nproblem: We introduce information-theoretic scores for privacy and utility,\nwhich quantify the average performance of an unfaithful user (e.g., adversary)\nand a faithful user (e.g., model developer) that have access to the published\nencoded data. We then theoretically characterize primitives in building\nfamilies of encoding schemes that motivate the use of random deep neural\nnetworks. Empirically, we compare the performance of our randomized encoding\nscheme and a linear scheme to a suite of computational attacks, and we also\nshow that our scheme achieves competitive prediction accuracy to raw-sample\nbaselines. Moreover, we demonstrate that multiple institutions, using\nindependent random encoders, can collaborate to train improved ML models.",
"title": "PEOPL: Characterizing Privately Encoded Open Datasets with Public Labels",
"url": "http://arxiv.org/abs/2304.00047v1"
} | null | null | no_new_dataset | admin | null | false | null | 159e2249-b5f9-4f5c-bced-d3c8106f4bc3 | null | Validated | {
"text_length": 1481
} | 1no_new_dataset
|
TITLE: Machine Learning Models Evaluation and Feature Importance Analysis on NPL Dataset
ABSTRACT: Predicting the probability of non-performing loans for individuals has a
vital and beneficial role for banks to decrease credit risk and make the right
decisions before giving the loan. The trend to make these decisions are based
on credit study and in accordance with generally accepted standards, loan
payment history, and demographic data of the clients. In this work, we evaluate
how different Machine learning models such as Random Forest, Decision tree,
KNN, SVM, and XGBoost perform on the dataset provided by a private bank in
Ethiopia. Further, motivated by this evaluation we explore different feature
selection methods to state the important features for the bank. Our findings
show that XGBoost achieves the highest F1 score on the KMeans SMOTE
over-sampled data. We also found that the most important features are the age
of the applicant, years of employment, and total income of the applicant rather
than collateral-related features in evaluating credit risk. | {
"abstract": "Predicting the probability of non-performing loans for individuals has a\nvital and beneficial role for banks to decrease credit risk and make the right\ndecisions before giving the loan. The trend to make these decisions are based\non credit study and in accordance with generally accepted standards, loan\npayment history, and demographic data of the clients. In this work, we evaluate\nhow different Machine learning models such as Random Forest, Decision tree,\nKNN, SVM, and XGBoost perform on the dataset provided by a private bank in\nEthiopia. Further, motivated by this evaluation we explore different feature\nselection methods to state the important features for the bank. Our findings\nshow that XGBoost achieves the highest F1 score on the KMeans SMOTE\nover-sampled data. We also found that the most important features are the age\nof the applicant, years of employment, and total income of the applicant rather\nthan collateral-related features in evaluating credit risk.",
"title": "Machine Learning Models Evaluation and Feature Importance Analysis on NPL Dataset",
"url": "http://arxiv.org/abs/2209.09638v1"
} | null | null | no_new_dataset | admin | null | false | null | f7f340d8-ba7b-4189-9251-379eb2f0a28b | null | Validated | {
"text_length": 1090
} | 1no_new_dataset
|
TITLE: WikiDes: A Wikipedia-Based Dataset for Generating Short Descriptions from Paragraphs
ABSTRACT: As free online encyclopedias with massive volumes of content, Wikipedia and
Wikidata are key to many Natural Language Processing (NLP) tasks, such as
information retrieval, knowledge base building, machine translation, text
classification, and text summarization. In this paper, we introduce WikiDes, a
novel dataset to generate short descriptions of Wikipedia articles for the
problem of text summarization. The dataset consists of over 80k English samples
on 6987 topics. We set up a two-phase summarization method - description
generation (Phase I) and candidate ranking (Phase II) - as a strong approach
that relies on transfer and contrastive learning. For description generation,
T5 and BART show their superiority compared to other small-scale pre-trained
models. By applying contrastive learning with the diverse input from beam
search, the metric fusion-based ranking models outperform the direct
description generation models significantly up to 22 ROUGE in topic-exclusive
split and topic-independent split. Furthermore, the outcome descriptions in
Phase II are supported by human evaluation in over 45.33% chosen compared to
23.66% in Phase I against the gold descriptions. In the aspect of sentiment
analysis, the generated descriptions cannot effectively capture all sentiment
polarities from paragraphs while doing this task better from the gold
descriptions. The automatic generation of new descriptions reduces the human
efforts in creating them and enriches Wikidata-based knowledge graphs. Our
paper shows a practical impact on Wikipedia and Wikidata since there are
thousands of missing descriptions. Finally, we expect WikiDes to be a useful
dataset for related works in capturing salient information from short
paragraphs. The curated dataset is publicly available at:
https://github.com/declare-lab/WikiDes. | {
"abstract": "As free online encyclopedias with massive volumes of content, Wikipedia and\nWikidata are key to many Natural Language Processing (NLP) tasks, such as\ninformation retrieval, knowledge base building, machine translation, text\nclassification, and text summarization. In this paper, we introduce WikiDes, a\nnovel dataset to generate short descriptions of Wikipedia articles for the\nproblem of text summarization. The dataset consists of over 80k English samples\non 6987 topics. We set up a two-phase summarization method - description\ngeneration (Phase I) and candidate ranking (Phase II) - as a strong approach\nthat relies on transfer and contrastive learning. For description generation,\nT5 and BART show their superiority compared to other small-scale pre-trained\nmodels. By applying contrastive learning with the diverse input from beam\nsearch, the metric fusion-based ranking models outperform the direct\ndescription generation models significantly up to 22 ROUGE in topic-exclusive\nsplit and topic-independent split. Furthermore, the outcome descriptions in\nPhase II are supported by human evaluation in over 45.33% chosen compared to\n23.66% in Phase I against the gold descriptions. In the aspect of sentiment\nanalysis, the generated descriptions cannot effectively capture all sentiment\npolarities from paragraphs while doing this task better from the gold\ndescriptions. The automatic generation of new descriptions reduces the human\nefforts in creating them and enriches Wikidata-based knowledge graphs. Our\npaper shows a practical impact on Wikipedia and Wikidata since there are\nthousands of missing descriptions. Finally, we expect WikiDes to be a useful\ndataset for related works in capturing salient information from short\nparagraphs. The curated dataset is publicly available at:\nhttps://github.com/declare-lab/WikiDes.",
"title": "WikiDes: A Wikipedia-Based Dataset for Generating Short Descriptions from Paragraphs",
"url": "http://arxiv.org/abs/2209.13101v1"
} | null | null | new_dataset | admin | null | false | null | 7b7d27a2-f818-42b4-8518-49a094b5cbad | null | Validated | {
"text_length": 1949
} | 0new_dataset
|
TITLE: Dark solitons in Bose-Einstein condensates: a dataset for many-body physics research
ABSTRACT: We establish a dataset of over $1.6\times10^4$ experimental images of
Bose--Einstein condensates containing solitonic excitations to enable machine
learning (ML) for many-body physics research. About $33~\%$ of this dataset has
manually assigned and carefully curated labels. The remainder is automatically
labeled using SolDet -- an implementation of a physics-informed ML data
analysis framework -- consisting of a convolutional-neural-network-based
classifier and OD as well as a statistically motivated physics-informed
classifier and a quality metric. This technical note constitutes the definitive
reference of the dataset, providing an opportunity for the data science
community to develop more sophisticated analysis tools, to further understand
nonlinear many-body physics, and even advance cold atom experiments. | {
"abstract": "We establish a dataset of over $1.6\\times10^4$ experimental images of\nBose--Einstein condensates containing solitonic excitations to enable machine\nlearning (ML) for many-body physics research. About $33~\\%$ of this dataset has\nmanually assigned and carefully curated labels. The remainder is automatically\nlabeled using SolDet -- an implementation of a physics-informed ML data\nanalysis framework -- consisting of a convolutional-neural-network-based\nclassifier and OD as well as a statistically motivated physics-informed\nclassifier and a quality metric. This technical note constitutes the definitive\nreference of the dataset, providing an opportunity for the data science\ncommunity to develop more sophisticated analysis tools, to further understand\nnonlinear many-body physics, and even advance cold atom experiments.",
"title": "Dark solitons in Bose-Einstein condensates: a dataset for many-body physics research",
"url": "http://arxiv.org/abs/2205.09114v2"
} | null | null | new_dataset | admin | null | false | null | da0b69b2-3d55-4aaf-9740-5392412c8e77 | null | Validated | {
"text_length": 941
} | 0new_dataset
|
TITLE: SubseasonalClimateUSA: A Dataset for Subseasonal Forecasting and Benchmarking
ABSTRACT: Subseasonal forecasting of the weather two to six weeks in advance is
critical for resource allocation and climate adaptation but poses many
challenges for the forecasting community. At this forecast horizon,
physics-based dynamical models have limited skill, and the targets for
prediction depend in a complex manner on both local weather and global climate
variables. Recently, machine learning methods have shown promise in advancing
the state of the art but only at the cost of complex data curation, integrating
expert knowledge with aggregation across multiple relevant data sources, file
formats, and temporal and spatial resolutions. To streamline this process and
accelerate future development, we introduce SubseasonalClimateUSA, a curated
dataset for training and benchmarking subseasonal forecasting models in the
United States. We use this dataset to benchmark a diverse suite of subseasonal
models, including operational dynamical models, classical meteorological
baselines, and ten state-of-the-art machine learning and deep learning-based
methods from the literature. Overall, our benchmarks suggest simple and
effective ways to extend the accuracy of current operational models.
SubseasonalClimateUSA is regularly updated and accessible via the
https://github.com/microsoft/subseasonal_data/ Python package. | {
"abstract": "Subseasonal forecasting of the weather two to six weeks in advance is\ncritical for resource allocation and climate adaptation but poses many\nchallenges for the forecasting community. At this forecast horizon,\nphysics-based dynamical models have limited skill, and the targets for\nprediction depend in a complex manner on both local weather and global climate\nvariables. Recently, machine learning methods have shown promise in advancing\nthe state of the art but only at the cost of complex data curation, integrating\nexpert knowledge with aggregation across multiple relevant data sources, file\nformats, and temporal and spatial resolutions. To streamline this process and\naccelerate future development, we introduce SubseasonalClimateUSA, a curated\ndataset for training and benchmarking subseasonal forecasting models in the\nUnited States. We use this dataset to benchmark a diverse suite of subseasonal\nmodels, including operational dynamical models, classical meteorological\nbaselines, and ten state-of-the-art machine learning and deep learning-based\nmethods from the literature. Overall, our benchmarks suggest simple and\neffective ways to extend the accuracy of current operational models.\nSubseasonalClimateUSA is regularly updated and accessible via the\nhttps://github.com/microsoft/subseasonal_data/ Python package.",
"title": "SubseasonalClimateUSA: A Dataset for Subseasonal Forecasting and Benchmarking",
"url": "http://arxiv.org/abs/2109.10399v3"
} | null | null | new_dataset | admin | null | false | null | 0bc818ba-e944-48fa-b660-12e49dde2661 | null | Validated | {
"text_length": 1436
} | 0new_dataset
|
TITLE: OLIVES Dataset: Ophthalmic Labels for Investigating Visual Eye Semantics
ABSTRACT: Clinical diagnosis of the eye is performed over multifarious data modalities
including scalar clinical labels, vectorized biomarkers, two-dimensional fundus
images, and three-dimensional Optical Coherence Tomography (OCT) scans.
Clinical practitioners use all available data modalities for diagnosing and
treating eye diseases like Diabetic Retinopathy (DR) or Diabetic Macular Edema
(DME). Enabling usage of machine learning algorithms within the ophthalmic
medical domain requires research into the relationships and interactions
between all relevant data over a treatment period. Existing datasets are
limited in that they neither provide data nor consider the explicit
relationship modeling between the data modalities. In this paper, we introduce
the Ophthalmic Labels for Investigating Visual Eye Semantics (OLIVES) dataset
that addresses the above limitation. This is the first OCT and near-IR fundus
dataset that includes clinical labels, biomarker labels, disease labels, and
time-series patient treatment information from associated clinical trials. The
dataset consists of 1268 near-IR fundus images each with at least 49 OCT scans,
and 16 biomarkers, along with 4 clinical labels and a disease diagnosis of DR
or DME. In total, there are 96 eyes' data averaged over a period of at least
two years with each eye treated for an average of 66 weeks and 7 injections. We
benchmark the utility of OLIVES dataset for ophthalmic data as well as provide
benchmarks and concrete research directions for core and emerging machine
learning paradigms within medical image analysis. | {
"abstract": "Clinical diagnosis of the eye is performed over multifarious data modalities\nincluding scalar clinical labels, vectorized biomarkers, two-dimensional fundus\nimages, and three-dimensional Optical Coherence Tomography (OCT) scans.\nClinical practitioners use all available data modalities for diagnosing and\ntreating eye diseases like Diabetic Retinopathy (DR) or Diabetic Macular Edema\n(DME). Enabling usage of machine learning algorithms within the ophthalmic\nmedical domain requires research into the relationships and interactions\nbetween all relevant data over a treatment period. Existing datasets are\nlimited in that they neither provide data nor consider the explicit\nrelationship modeling between the data modalities. In this paper, we introduce\nthe Ophthalmic Labels for Investigating Visual Eye Semantics (OLIVES) dataset\nthat addresses the above limitation. This is the first OCT and near-IR fundus\ndataset that includes clinical labels, biomarker labels, disease labels, and\ntime-series patient treatment information from associated clinical trials. The\ndataset consists of 1268 near-IR fundus images each with at least 49 OCT scans,\nand 16 biomarkers, along with 4 clinical labels and a disease diagnosis of DR\nor DME. In total, there are 96 eyes' data averaged over a period of at least\ntwo years with each eye treated for an average of 66 weeks and 7 injections. We\nbenchmark the utility of OLIVES dataset for ophthalmic data as well as provide\nbenchmarks and concrete research directions for core and emerging machine\nlearning paradigms within medical image analysis.",
"title": "OLIVES Dataset: Ophthalmic Labels for Investigating Visual Eye Semantics",
"url": "http://arxiv.org/abs/2209.11195v1"
} | null | null | new_dataset | admin | null | false | null | af34e045-9e03-465b-8614-f26ff70ba6d1 | null | Validated | {
"text_length": 1688
} | 0new_dataset
|
TITLE: Making Machine Learning Datasets and Models FAIR for HPC: A Methodology and Case Study
ABSTRACT: The FAIR Guiding Principles aim to improve the findability, accessibility,
interoperability, and reusability of digital content by making them both human
and machine actionable. However, these principles have not yet been broadly
adopted in the domain of machine learning-based program analyses and
optimizations for High-Performance Computing (HPC). In this paper, we design a
methodology to make HPC datasets and machine learning models FAIR after
investigating existing FAIRness assessment and improvement techniques. Our
methodology includes a comprehensive, quantitative assessment for elected data,
followed by concrete, actionable suggestions to improve FAIRness with respect
to common issues related to persistent identifiers, rich metadata descriptions,
license and provenance information. Moreover, we select a representative
training dataset to evaluate our methodology. The experiment shows the
methodology can effectively improve the dataset and model's FAIRness from an
initial score of 19.1% to the final score of 83.0%. | {
"abstract": "The FAIR Guiding Principles aim to improve the findability, accessibility,\ninteroperability, and reusability of digital content by making them both human\nand machine actionable. However, these principles have not yet been broadly\nadopted in the domain of machine learning-based program analyses and\noptimizations for High-Performance Computing (HPC). In this paper, we design a\nmethodology to make HPC datasets and machine learning models FAIR after\ninvestigating existing FAIRness assessment and improvement techniques. Our\nmethodology includes a comprehensive, quantitative assessment for elected data,\nfollowed by concrete, actionable suggestions to improve FAIRness with respect\nto common issues related to persistent identifiers, rich metadata descriptions,\nlicense and provenance information. Moreover, we select a representative\ntraining dataset to evaluate our methodology. The experiment shows the\nmethodology can effectively improve the dataset and model's FAIRness from an\ninitial score of 19.1% to the final score of 83.0%.",
"title": "Making Machine Learning Datasets and Models FAIR for HPC: A Methodology and Case Study",
"url": "http://arxiv.org/abs/2211.02092v1"
} | null | null | no_new_dataset | admin | null | false | null | 0e687961-6e1d-4894-976e-690d0f572f05 | null | Validated | {
"text_length": 1156
} | 1no_new_dataset
|
TITLE: MangoLeafBD: A Comprehensive Image Dataset to Classify Diseased and Healthy Mango Leaves
ABSTRACT: Agriculture is of one of the few remaining sectors that is yet to receive
proper attention from the machine learning community. The importance of
datasets in the machine learning discipline cannot be overemphasized. The lack
of standard and publicly available datasets related to agriculture impedes
practitioners of this discipline to harness the full benefit of these powerful
computational predictive tools and techniques. To improve this scenario, we
develop, to the best of our knowledge, the first-ever standard, ready-to-use,
and publicly available dataset of mango leaves. The images are collected from
four mango orchards of Bangladesh, one of the top mango-growing countries of
the world. The dataset contains 4000 images of about 1800 distinct leaves
covering seven diseases. Although the dataset is developed using mango leaves
of Bangladesh only, since we deal with diseases that are common across many
countries, this dataset is likely to be applicable to identify mango diseases
in other countries as well, thereby boosting mango yield. This dataset is
expected to draw wide attention from machine learning researchers and
practitioners in the field of automated agriculture. | {
"abstract": "Agriculture is of one of the few remaining sectors that is yet to receive\nproper attention from the machine learning community. The importance of\ndatasets in the machine learning discipline cannot be overemphasized. The lack\nof standard and publicly available datasets related to agriculture impedes\npractitioners of this discipline to harness the full benefit of these powerful\ncomputational predictive tools and techniques. To improve this scenario, we\ndevelop, to the best of our knowledge, the first-ever standard, ready-to-use,\nand publicly available dataset of mango leaves. The images are collected from\nfour mango orchards of Bangladesh, one of the top mango-growing countries of\nthe world. The dataset contains 4000 images of about 1800 distinct leaves\ncovering seven diseases. Although the dataset is developed using mango leaves\nof Bangladesh only, since we deal with diseases that are common across many\ncountries, this dataset is likely to be applicable to identify mango diseases\nin other countries as well, thereby boosting mango yield. This dataset is\nexpected to draw wide attention from machine learning researchers and\npractitioners in the field of automated agriculture.",
"title": "MangoLeafBD: A Comprehensive Image Dataset to Classify Diseased and Healthy Mango Leaves",
"url": "http://arxiv.org/abs/2209.02377v1"
} | null | null | new_dataset | admin | null | false | null | ee1b1522-81b9-44c6-a39f-d3279b89af1c | null | Validated | {
"text_length": 1313
} | 0new_dataset
|
TITLE: DuReader_robust: A Chinese Dataset Towards Evaluating Robustness and Generalization of Machine Reading Comprehension in Real-World Applications
ABSTRACT: Machine reading comprehension (MRC) is a crucial task in natural language
processing and has achieved remarkable advancements. However, most of the
neural MRC models are still far from robust and fail to generalize well in
real-world applications. In order to comprehensively verify the robustness and
generalization of MRC models, we introduce a real-world Chinese dataset --
DuReader_robust. It is designed to evaluate the MRC models from three aspects:
over-sensitivity, over-stability and generalization. Comparing to previous
work, the instances in DuReader_robust are natural texts, rather than the
altered unnatural texts. It presents the challenges when applying MRC models to
real-world applications. The experimental results show that MRC models do not
perform well on the challenge test set. Moreover, we analyze the behavior of
existing models on the challenge test set, which may provide suggestions for
future model development. The dataset and codes are publicly available at
https://github.com/baidu/DuReader. | {
"abstract": "Machine reading comprehension (MRC) is a crucial task in natural language\nprocessing and has achieved remarkable advancements. However, most of the\nneural MRC models are still far from robust and fail to generalize well in\nreal-world applications. In order to comprehensively verify the robustness and\ngeneralization of MRC models, we introduce a real-world Chinese dataset --\nDuReader_robust. It is designed to evaluate the MRC models from three aspects:\nover-sensitivity, over-stability and generalization. Comparing to previous\nwork, the instances in DuReader_robust are natural texts, rather than the\naltered unnatural texts. It presents the challenges when applying MRC models to\nreal-world applications. The experimental results show that MRC models do not\nperform well on the challenge test set. Moreover, we analyze the behavior of\nexisting models on the challenge test set, which may provide suggestions for\nfuture model development. The dataset and codes are publicly available at\nhttps://github.com/baidu/DuReader.",
"title": "DuReader_robust: A Chinese Dataset Towards Evaluating Robustness and Generalization of Machine Reading Comprehension in Real-World Applications",
"url": "http://arxiv.org/abs/2004.11142v2"
} | null | null | new_dataset | admin | null | false | null | 1328354b-0919-4070-949f-efbc0212e99f | null | Validated | {
"text_length": 1203
} | 0new_dataset
|
TITLE: Wide-scale Monitoring of Satellite Lifetimes: Pitfalls and a Benchmark Dataset
ABSTRACT: An important task within the broader goal of Space Situational Awareness
(SSA) is to observe changes in the orbits of satellites, where the data spans
thousands of objects over long time scales (decades). The Two-Line Element
(TLE) data provided by the North American Aerospace Defense Command is the most
comprehensive and widely-available dataset cataloguing the orbits of
satellites. This makes it a highly-attractive data source on which to perform
this observation. However, when attempting to infer changes in satellite
behaviour from TLE data, there are a number of potential pitfalls. These mostly
relate to specific features of the TLE data which are not always clearly
documented in the data sources or popular software packages for manipulating
them. These quirks produce a particularly hazardous data type for researchers
from adjacent disciplines (such as anomaly detection or machine learning). We
highlight these features of TLE data and the resulting pitfalls in order to
save future researchers from being trapped. A seperate, significant, issue is
that existing contributions to manoeuvre detection from TLE data evaluate their
algorithms on different satellites, making comparison between these methods
difficult. Moreover, the ground-truth in these datasets is often poor quality,
sometimes being based on subjective human assessment. We therefore release and
describe in-depth an open, curated, benchmark dataset containing TLE data for
15 satellites alongside high-quality ground-truth manoeuvre timestamps. | {
"abstract": "An important task within the broader goal of Space Situational Awareness\n(SSA) is to observe changes in the orbits of satellites, where the data spans\nthousands of objects over long time scales (decades). The Two-Line Element\n(TLE) data provided by the North American Aerospace Defense Command is the most\ncomprehensive and widely-available dataset cataloguing the orbits of\nsatellites. This makes it a highly-attractive data source on which to perform\nthis observation. However, when attempting to infer changes in satellite\nbehaviour from TLE data, there are a number of potential pitfalls. These mostly\nrelate to specific features of the TLE data which are not always clearly\ndocumented in the data sources or popular software packages for manipulating\nthem. These quirks produce a particularly hazardous data type for researchers\nfrom adjacent disciplines (such as anomaly detection or machine learning). We\nhighlight these features of TLE data and the resulting pitfalls in order to\nsave future researchers from being trapped. A seperate, significant, issue is\nthat existing contributions to manoeuvre detection from TLE data evaluate their\nalgorithms on different satellites, making comparison between these methods\ndifficult. Moreover, the ground-truth in these datasets is often poor quality,\nsometimes being based on subjective human assessment. We therefore release and\ndescribe in-depth an open, curated, benchmark dataset containing TLE data for\n15 satellites alongside high-quality ground-truth manoeuvre timestamps.",
"title": "Wide-scale Monitoring of Satellite Lifetimes: Pitfalls and a Benchmark Dataset",
"url": "http://arxiv.org/abs/2212.08662v1"
} | null | null | new_dataset | admin | null | false | null | 835672b4-b22c-483a-8ce2-db7d4f2ea2b0 | null | Validated | {
"text_length": 1642
} | 0new_dataset
|
TITLE: BiRdQA: A Bilingual Dataset for Question Answering on Tricky Riddles
ABSTRACT: A riddle is a question or statement with double or veiled meanings, followed
by an unexpected answer. Solving riddle is a challenging task for both machine
and human, testing the capability of understanding figurative, creative natural
language and reasoning with commonsense knowledge. We introduce BiRdQA, a
bilingual multiple-choice question answering dataset with 6614 English riddles
and 8751 Chinese riddles. For each riddle-answer pair, we provide four
distractors with additional information from Wikipedia. The distractors are
automatically generated at scale with minimal bias. Existing monolingual and
multilingual QA models fail to perform well on our dataset, indicating that
there is a long way to go before machine can beat human on solving tricky
riddles. The dataset has been released to the community. | {
"abstract": "A riddle is a question or statement with double or veiled meanings, followed\nby an unexpected answer. Solving riddle is a challenging task for both machine\nand human, testing the capability of understanding figurative, creative natural\nlanguage and reasoning with commonsense knowledge. We introduce BiRdQA, a\nbilingual multiple-choice question answering dataset with 6614 English riddles\nand 8751 Chinese riddles. For each riddle-answer pair, we provide four\ndistractors with additional information from Wikipedia. The distractors are\nautomatically generated at scale with minimal bias. Existing monolingual and\nmultilingual QA models fail to perform well on our dataset, indicating that\nthere is a long way to go before machine can beat human on solving tricky\nriddles. The dataset has been released to the community.",
"title": "BiRdQA: A Bilingual Dataset for Question Answering on Tricky Riddles",
"url": "http://arxiv.org/abs/2109.11087v2"
} | null | null | new_dataset | admin | null | false | null | 039a5a67-2e67-40ac-8b05-939db7e0d062 | null | Validated | {
"text_length": 922
} | 0new_dataset
|
TITLE: Anomaly Detection and Inter-Sensor Transfer Learning on Smart Manufacturing Datasets
ABSTRACT: Smart manufacturing systems are being deployed at a growing rate because of
their ability to interpret a wide variety of sensed information and act on the
knowledge gleaned from system observations. In many cases, the principal goal
of the smart manufacturing system is to rapidly detect (or anticipate) failures
to reduce operational cost and eliminate downtime. This often boils down to
detecting anomalies within the sensor date acquired from the system. The smart
manufacturing application domain poses certain salient technical challenges. In
particular, there are often multiple types of sensors with varying capabilities
and costs. The sensor data characteristics change with the operating point of
the environment or machines, such as, the RPM of the motor. The anomaly
detection process therefore has to be calibrated near an operating point. In
this paper, we analyze four datasets from sensors deployed from manufacturing
testbeds. We evaluate the performance of several traditional and ML-based
forecasting models for predicting the time series of sensor data. Then,
considering the sparse data from one kind of sensor, we perform transfer
learning from a high data rate sensor to perform defect type classification.
Taken together, we show that predictive failure classification can be achieved,
thus paving the way for predictive maintenance. | {
"abstract": "Smart manufacturing systems are being deployed at a growing rate because of\ntheir ability to interpret a wide variety of sensed information and act on the\nknowledge gleaned from system observations. In many cases, the principal goal\nof the smart manufacturing system is to rapidly detect (or anticipate) failures\nto reduce operational cost and eliminate downtime. This often boils down to\ndetecting anomalies within the sensor date acquired from the system. The smart\nmanufacturing application domain poses certain salient technical challenges. In\nparticular, there are often multiple types of sensors with varying capabilities\nand costs. The sensor data characteristics change with the operating point of\nthe environment or machines, such as, the RPM of the motor. The anomaly\ndetection process therefore has to be calibrated near an operating point. In\nthis paper, we analyze four datasets from sensors deployed from manufacturing\ntestbeds. We evaluate the performance of several traditional and ML-based\nforecasting models for predicting the time series of sensor data. Then,\nconsidering the sparse data from one kind of sensor, we perform transfer\nlearning from a high data rate sensor to perform defect type classification.\nTaken together, we show that predictive failure classification can be achieved,\nthus paving the way for predictive maintenance.",
"title": "Anomaly Detection and Inter-Sensor Transfer Learning on Smart Manufacturing Datasets",
"url": "http://arxiv.org/abs/2206.06355v1"
} | null | null | no_new_dataset | admin | null | false | null | 87794a9a-f4f3-44b5-9795-9c7573fa088d | null | Validated | {
"text_length": 1475
} | 1no_new_dataset
|
TITLE: Multimodal Lecture Presentations Dataset: Understanding Multimodality in Educational Slides
ABSTRACT: Lecture slide presentations, a sequence of pages that contain text and
figures accompanied by speech, are constructed and presented carefully in order
to optimally transfer knowledge to students. Previous studies in multimedia and
psychology attribute the effectiveness of lecture presentations to their
multimodal nature. As a step toward developing AI to aid in student learning as
intelligent teacher assistants, we introduce the Multimodal Lecture
Presentations dataset as a large-scale benchmark testing the capabilities of
machine learning models in multimodal understanding of educational content. Our
dataset contains aligned slides and spoken language, for 180+ hours of video
and 9000+ slides, with 10 lecturers from various subjects (e.g., computer
science, dentistry, biology). We introduce two research tasks which are
designed as stepping stones towards AI agents that can explain (automatically
captioning a lecture presentation) and illustrate (synthesizing visual figures
to accompany spoken explanations) educational content. We provide manual
annotations to help implement these two research tasks and evaluate
state-of-the-art models on them. Comparing baselines and human student
performances, we find that current models struggle in (1) weak crossmodal
alignment between slides and spoken text, (2) learning novel visual mediums,
(3) technical language, and (4) long-range sequences. Towards addressing this
issue, we also introduce PolyViLT, a multimodal transformer trained with a
multi-instance learning loss that is more effective than current approaches. We
conclude by shedding light on the challenges and opportunities in multimodal
understanding of educational presentations. | {
"abstract": "Lecture slide presentations, a sequence of pages that contain text and\nfigures accompanied by speech, are constructed and presented carefully in order\nto optimally transfer knowledge to students. Previous studies in multimedia and\npsychology attribute the effectiveness of lecture presentations to their\nmultimodal nature. As a step toward developing AI to aid in student learning as\nintelligent teacher assistants, we introduce the Multimodal Lecture\nPresentations dataset as a large-scale benchmark testing the capabilities of\nmachine learning models in multimodal understanding of educational content. Our\ndataset contains aligned slides and spoken language, for 180+ hours of video\nand 9000+ slides, with 10 lecturers from various subjects (e.g., computer\nscience, dentistry, biology). We introduce two research tasks which are\ndesigned as stepping stones towards AI agents that can explain (automatically\ncaptioning a lecture presentation) and illustrate (synthesizing visual figures\nto accompany spoken explanations) educational content. We provide manual\nannotations to help implement these two research tasks and evaluate\nstate-of-the-art models on them. Comparing baselines and human student\nperformances, we find that current models struggle in (1) weak crossmodal\nalignment between slides and spoken text, (2) learning novel visual mediums,\n(3) technical language, and (4) long-range sequences. Towards addressing this\nissue, we also introduce PolyViLT, a multimodal transformer trained with a\nmulti-instance learning loss that is more effective than current approaches. We\nconclude by shedding light on the challenges and opportunities in multimodal\nunderstanding of educational presentations.",
"title": "Multimodal Lecture Presentations Dataset: Understanding Multimodality in Educational Slides",
"url": "http://arxiv.org/abs/2208.08080v1"
} | null | null | new_dataset | admin | null | false | null | 1d4f2924-deb2-4c82-9a01-95e659100428 | null | Validated | {
"text_length": 1831
} | 0new_dataset
|
TITLE: On the Robustness of Dataset Inference
ABSTRACT: Machine learning (ML) models are costly to train as they can require a
significant amount of data, computational resources and technical expertise.
Thus, they constitute valuable intellectual property that needs protection from
adversaries wanting to steal them. Ownership verification techniques allow the
victims of model stealing attacks to demonstrate that a suspect model was in
fact stolen from theirs.
Although a number of ownership verification techniques based on watermarking
or fingerprinting have been proposed, most of them fall short either in terms
of security guarantees (well-equipped adversaries can evade verification) or
computational cost. A fingerprinting technique, Dataset Inference (DI), has
been shown to offer better robustness and efficiency than prior methods.
The authors of DI provided a correctness proof for linear (suspect) models.
However, in a subspace of the same setting, we prove that DI suffers from high
false positives (FPs) -- it can incorrectly identify an independent model
trained with non-overlapping data from the same distribution as stolen. We
further prove that DI also triggers FPs in realistic, non-linear suspect
models. We then confirm empirically that DI in the black-box setting leads to
FPs, with high confidence.
Second, we show that DI also suffers from false negatives (FNs) -- an
adversary can fool DI (at the cost of incurring some accuracy loss) by
regularising a stolen model's decision boundaries using adversarial training,
thereby leading to an FN. To this end, we demonstrate that black-box DI fails
to identify a model adversarially trained from a stolen dataset -- the setting
where DI is the hardest to evade.
Finally, we discuss the implications of our findings, the viability of
fingerprinting-based ownership verification in general, and suggest directions
for future work. | {
"abstract": "Machine learning (ML) models are costly to train as they can require a\nsignificant amount of data, computational resources and technical expertise.\nThus, they constitute valuable intellectual property that needs protection from\nadversaries wanting to steal them. Ownership verification techniques allow the\nvictims of model stealing attacks to demonstrate that a suspect model was in\nfact stolen from theirs.\n Although a number of ownership verification techniques based on watermarking\nor fingerprinting have been proposed, most of them fall short either in terms\nof security guarantees (well-equipped adversaries can evade verification) or\ncomputational cost. A fingerprinting technique, Dataset Inference (DI), has\nbeen shown to offer better robustness and efficiency than prior methods.\n The authors of DI provided a correctness proof for linear (suspect) models.\nHowever, in a subspace of the same setting, we prove that DI suffers from high\nfalse positives (FPs) -- it can incorrectly identify an independent model\ntrained with non-overlapping data from the same distribution as stolen. We\nfurther prove that DI also triggers FPs in realistic, non-linear suspect\nmodels. We then confirm empirically that DI in the black-box setting leads to\nFPs, with high confidence.\n Second, we show that DI also suffers from false negatives (FNs) -- an\nadversary can fool DI (at the cost of incurring some accuracy loss) by\nregularising a stolen model's decision boundaries using adversarial training,\nthereby leading to an FN. To this end, we demonstrate that black-box DI fails\nto identify a model adversarially trained from a stolen dataset -- the setting\nwhere DI is the hardest to evade.\n Finally, we discuss the implications of our findings, the viability of\nfingerprinting-based ownership verification in general, and suggest directions\nfor future work.",
"title": "On the Robustness of Dataset Inference",
"url": "http://arxiv.org/abs/2210.13631v3"
} | null | null | no_new_dataset | admin | null | false | null | 0efab7b7-eeb5-4d56-93e5-6d4aa39164d8 | null | Validated | {
"text_length": 1929
} | 1no_new_dataset
|