MIND: A Large-scale Dataset For News Recommendation

1y ago
6 Views
1 Downloads
2.22 MB
10 Pages
Last View : 2m ago
Last Download : 3m ago
Upload by : Ryan Jay
Transcription

MIND: A Large-scale Dataset for News RecommendationFangzhao Wu† , Ying Qiao‡ , Jiun-Hung Chen‡ , Chuhan Wu§ , Tao Qi§ ,Jianxun Lian† , Danyang Liu† , Xing Xie† , Jianfeng Gao† , Winnie Wu‡ , Ming Zhou††‡§Microsoft ResearchMicrosoftTsinghua University{fangzwu, yiqia, jiuche, jialia}@microsoft.com{t-danliu, xingx, jfgao, winniew, mingzhou}@microsoft.com{wu-ch19, qit16}@mails.tsinghua.edu.cnAbstractNews recommendation is an important technique for personalized news service. Compared with product and movie recommendations which have been comprehensively studied, the research on news recommendation ismuch more limited, mainly due to the lack of ahigh-quality benchmark dataset. In this paper,we present a large-scale dataset named MINDfor news recommendation. Constructed fromthe user click logs of Microsoft News, MINDcontains 1 million users and more than 160kEnglish news articles, each of which has richtextual content such as title, abstract and body.We demonstrate MIND a good testbed fornews recommendation through a comparativestudy of several state-of-the-art news recommendation methods which are originally developed on different proprietary datasets. Ourresults show the performance of news recommendation highly relies on the quality of newscontent understanding and user interest modeling. Many natural language processing techniques such as effective text representationmethods and pre-trained language models caneffectively improve the performance of newsrecommendation. The MIND dataset will beavailable at https://msnews.github.io.1IntroductionOnline news services such as Google News and Microsoft News have become important platforms fora large population of users to obtain news information (Das et al., 2007; Wu et al., 2019a). Massivenews articles are generated and posted online everyday, making it difficult for users to find interestednews quickly (Okura et al., 2017). Personalizednews recommendation can help users alleviate information overload and improve news reading experience (Wu et al., 2019b). Thus, it is widely usedin many online news platforms (Li et al., 2011;Okura et al., 2017; An et al., 2019).In traditional recommender systems, users anditems are usually represented using IDs, and theirinteractions such as rating scores are used to learnID representations via methods like collaborativefiltering (Koren, 2008). However, news recommendation has some special challenges. First, newsarticles on news websites update very quickly. Newnews articles are posted continuously, and existingnews articles will expire in short time (Das et al.,2007). Thus, the cold-start problem is very severein news recommendation. Second, news articlescontain rich textual information such as title andbody. It is not appropriate to simply representing them using IDs, and it is important to understand their content from their texts (Kompan andBieliková, 2010). Third, there is no explicit ratingof news articles posted by users on news platforms.Thus, in news recommendation users’ interest innews is usually inferred from their click behaviorsin an implicit way (Ilievski and Roy, 2013).A large-scale and high-quality dataset can significantly facilitate the research in an area, suchas ImageNet for image classification (Deng et al.,2009) and SQuAD for machine reading comprehension (Rajpurkar et al., 2016). There are several public datasets for traditional recommendationtasks, such as Amazon dataset1 for product recommendation and MovieLens dataset2 for movierecommendation. Based on these datasets, manywell-known recommendation methods have beendeveloped. However, existing studies on news recommendation are much fewer, and many of themare conducted on proprietary datasets (Okura et al.,2017; Wang et al., 2018; Wu et al., 2019a). Although there are a few public datasets for newsrecommendation, they are usually in small size andmost of them are not in English. Thus, a ://grouplens.org/datasets/movielens/

Title(a) An example Microsoft News homepageMike Tomlin: Steelers ‘accept responsibility’ for role inbrawl with BrownsCategorySportsAbstractMike Tomlin has admitted that the Pittsburgh Steelersplayed a role in the brawl with the Cleveland Brownslast week, and on Tuesday he accepted responsibilityfor it on behalf of the organization.BodyTomlin opened his weekly news conference byaddressing the issue head on.“It was ugly,” said Tomlin, who had refused to take anyquestions about the incident directly after the game,per Brooke Pryor of ESPN. “It was ugly for the game offootball. I think all of us that are involved in the game,particularly at this level, (b) Texts in an example news articleFigure 1: An example homepage of Microsoft News and an example news article on it.large-scale English news recommendation datasetis of great value for the research in this area.In this paper we present a large-scale MIcrosoftNews Dataset (MIND) for news recommendationresearch, which is collected from the user behaviorlogs of Microsoft News3 . It contains 1 million usersand their click behaviors on more than 160k English news articles. We implement many state-ofthe-art news recommendation methods originallydeveloped on different proprietary datasets, andcompare their performance on the MIND datasetto provide a benchmark for news recommendationresearch. The experimental results show that adeep understanding of news articles through NLPtechniques is very important for news recommendation. Both effective text representation methodsand pre-trained language models can contribute tothe performance improvement of news recommendation. In addition, appropriate modeling of userinterest is also useful. We hope MIND can serveas a benchmark dataset for news recommendationand facilitate the research in this area.2Related Work2.1News RecommendationNews recommendation aims to find news articlesthat users have interest to read from the massivecandidate news (Das et al., 2007). There are twoimportant problems in news recommendation, i.e.,how to represent news articles which have rich textual content and how to model users’ interest innews from their previous behaviors (Okura et al.,2017). Traditional news recommendation methods usually rely on feature engineering to representnews articles and user interest (Liu et al., 2010;3https://microsoftnews.msn.com/Son et al., 2013; Karkali et al., 2013; Garcin et al.,2013; Bansal et al., 2015; Chen et al., 2017). Forexample, Li et al. (2010) represented news articlesusing their URLs and categories, and representedusers using their demographics, geographic information and behavior categories inferred from theirconsumption records on Yahoo!.In recent years, several deep learning based newsrecommendation methods have been proposed tolearn representations of news articles and user interest in an end-to-end manner (Okura et al., 2017;Wu et al., 2019a; An et al., 2019). For example, Okura et al. (2017) represented news articlesfrom news content using denoising autoencodermodel, and represented user interest from historicalclicked news articles with GRU model. Their experiments on Yahoo! Japan platform show that thenews and user representations learned with deeplearning models are promising for news recommendation. Wang et al. (2018) proposed to learnknowledge-aware news representations from newstitles using CNN network by incorporating bothword embeddings and the entity embeddings inferred from knowledge graphs. Wu et al. (2019a)proposed an attentive multi-view learning framework to represent news articles from different newstexts such as title, body and category. They usedan attention model to infer the interest of usersfrom their clicked news articles by selecting informative ones. These works are usually developedand validated on proprietary datasets which are notpublicly available, making it difficult for other researchers to verify these methods and develop theirown methods.News recommendation has rich inherent relatedness with natural language processing. First,news is a common form of texts, and text modeling

orwegianPortugueseEnglishEnglish# UsersUnknown3,083,438314,000Unknown1,000,000# News70,35348,48646,00014,180161,013# News informationtitle, bodytitle, body, categoryno original text, only word embeddingsno original text, only word IDstitle, abstract, body, categoryTable 1: Comparisons of the MIND dataset and the existing public news recommendation datasets.techniques such as CNN and Transformer can benaturally applied to represent news articles (Wuet al., 2019a; Ge et al., 2020). Second, learninguser interest representation from previously clickednews articles has similarity with learning documentrepresentation from its sentences. Third, newsrecommendation can be formulated as a specialtext matching problem, i.e., the matching betweena candidate news article and a set of previouslyclicked news articles in some news reading interestspace. Thus, news recommendation has attractedincreasing attentions from the NLP community (Anet al., 2019; Wu et al., 2019c).2.2Existing DatasetsThere are only a few public datasets for news recommendation, which are summarized in Table 1.Kille et al. (2013) constructed the Plista4 datasetby collecting news articles published on 13 German news portals and users’ click logs on them. Itcontains 70,353 news articles and 1,095,323 clickevents. The news articles in this dataset are inGerman and the users are mainly from the Germanspeaking world. Gulla et al. (2017) released theAdressa dataset5 , which was constructed from thelogs of the Adresseavisen website in ten weeks.It has 48,486 news articles, 3,083,438 users and27,223,576 click events. Each click event containsseveral features, such as session time, news title,news category and user ID. Each news article isassociated with some detailed information such asauthors, entities and body. The news articles inthis dataset are in Norwegian. Moreira et al. (2018)constructed a news recommendation dataset6 fromGlobo.com, a popular news portal in Brazil. Thisdataset contains about 314,000 users, 46,000 newsarticles and 3 million click records. Each clickrecord contains fields like user ID, news ID andsession time. Each news article has ID, category,publisher, creation time, and the embeddings ofits words generated by a neural model pre-trainedon a news metadata classification task (de SouzaPereira Moreira et al., 2018). However, the originaltexts of news articles are not provided. In addition,this dataset is in Portuguese. There is a Yahoo!dataset7 for session-based news recommendation.It contains 14,180 news articles and 34,022 clickevents. Each news article is represented by wordIDs, and the original news text is not provided.The number of users in this dataset is unknownsince there is no user ID. In summary, most existing public datasets for news recommendation arenon-English, and some of them are small in sizeand lack original news texts. Thus, a high-qualityEnglish news recommendation dataset is of greatvalue to the news recommendation community.3MIND Dataset3.1In order to facilitate the research in news recommendation, we built the MIcrosoft News Dataset(MIND)8 . It was collected from the user behaviorlogs of Microsoft News9 . We randomly sampled 1million users who had at least 5 news click recordsduring 6 weeks from October 12 to November 22,2019. In order to protect user privacy, each userwas de-linked from the production system whensecurely hashed into an anonymized ID using onetime salt10 mapping. We collected the behaviorlogs of these users in this period, which are formatted into impression logs. An impression logrecords the news articles displayed to a user whenshe visits the news website homepage at a specifictime, and her click behaviors on these news articles. Since in news recommendation we usuallypredict whether a user will click a candidate p?datatype lIt is public available at https://msnews.github.io for research purpose. Any question about this dataset can be sent m10https://en.wikipedia.org/wiki/Salt ions-by-globocom5Dataset Construction

article or not based on her personal interest inferredfrom her previous behaviors, we add the news clickhistories of users to their impression logs to construct labeled samples for training and verifyingnews recommendation models. The format of eachlabeled sample is [uID, t, ClickHist, ImpLog],where uID is the anonymous ID of a user, and tis the timestamp of this impression. ClickHist isan ID list of the news articles previously clicked bythis user (sorted by click time). ImpLog containsthe IDs of the news articles displayed in this impression and the labels indicating whether they areclicked, i.e., [(nID1 , label1 ), (nID2 , label2 ), .],where nID is news article ID and label is the clicklabel (1 for click and 0 for non-click). We usedthe samples in the last week for test, and the samples in the fifth week for training. For samples intraining set, we used the click behaviors in the firstfour weeks to construct the news click history. Forsamples in test set, the time period for news clickhistory extraction is the first five weeks. We onlykept the samples with non-empty news click history. Among the training data, we used the samplesin the last day of the fifth week as validation set.Each news article in the MIND dataset contains anews ID, a title, an abstract, a body, and a categorylabel such as “Sports” which is manually taggedby the editors. In addition, we found that thesenews texts contain rich entities. For example, inthe title of the news article shown in Fig. 1 “MikeTomlin: Steelers ‘accept responsibility’ for role inbrawl with Browns”, “Mike Tomlin” is a personentity, and “Steelers” and “Browns” are entities ofAmerican football team. In order to facilitate the research of knowledge-aware news recommendation,we extracted the entities in the titles, abstracts andbodies of the news articles in the MIND dataset,and linked them to the entities in WikiData11 using an internal NER and entity linking tool. Wealso extracted the knowledge triples of these entities from WikiData and used TransE (Bordes et al.,2013) method to learn the embeddings of entitiesand relations. These entities, knowledge triples,as well as entity and relation embeddings are alsoincluded in the MIND dataset.3.2Dataset AnalysisThe detailed statistics of the MIND dataset are summarized in Table 2 and Fig. 2. This dataset contains1,000,000 users and 161,013 news articles. Page(a) Title Length(b) Abstract Length(c) Body Length(d) Survival TimeFigure 2: Key statistics of the MIND dataset.# News# News category# EntityAvg. title len.Avg. body len.161,013203,299,68711.52585.05# Users# Impression# Click behaviorAvg. abstract len.1,000,00015,777,37724,155,47043.00Table 2: Detailed statistics of the MIND dataset.are 2,186,683 samples in the training set, 365,200samples in the validation set, and 2,341,619 samples in the test set, which can empower the trainingof data-intensive news recommendation models.Figs. 2(a), 2(b) and 2(c) show the length distributions of news title, abstract and body. We can seethat news titles are usually very short and the average length is only 11.52 words. In comparison,news abstracts and bodies are much longer andmay contain richer information of news content.Thus, incorporating different kinds of news information such as title, abstract and body may helpunderstand news articles better.Fig. 2(d) shows the survival time distribution ofnews articles. The survival time of a news articleis estimated here using the time interval betweenits first and last appearance time in the dataset. Wefind that the survival time of more than 84.5% newsarticles is less than two days. This is due to the nature of news information, since news media alwayspursue the latest news and the exiting news articlesget out-of-date quickly. Thus, cold-start problemis a common phenomenon in news recommendation, and the traditional ID-based recommendersystems (Koren, 2008) are not suitable for this task.Representing news articles using their textual content is critical for news recommendation.

4MethodIn this section, we briefly introduce several methods for news recommendation, including generalrecommendation methods and news-specific recommendation methods. These methods were developed in different settings and on different datasets.Some of their implementations can be found in Microsoft Recommenders open source repository12 .We will compare them on the MIND dataset.4.1General Recommendation MethodsLibFM (Rendle, 2012), a classic recommendationmethod based on factorization machine. Besidesthe user ID and news ID, we also use the contentfeatures13 extracted from previously clicked newsand candidate news as the additional features torepresent users and candidate news.DSSM (Huang et al., 2013), deep structured semantic model, which uses tri-gram hashes and multiplefeed-forward neural networks for query-documentmatching. We use the content features extractedfrom previous clicked news as query, and thosefrom candidate news as document.Wide&Deep (Cheng et al., 2016), a two-channelneural recommendation method, which has a widelinear transformation channel and a deep neuralnetwork channel. We use the same content featuresof users and candidate news for both channels.DeepFM (Guo et al., 2017), another popular neuralrecommendation method which synthesizes deepneural networks and factorization machines. Thesame content features of users and candidate newsare fed to both components.4.2News Recommendation MethodsDFM (Lian et al., 2018), deep fusion model, a newsrecommendation method which uses an inceptionnetwork to combine neural networks with differentdepths to capture the complex interactions betweenfeatures. We use the same features of users andcandidate news with aforementioned methods.GRU (Okura et al., 2017), a neural news recommendation method which uses autoencoder to learnlatent news representations from news content, anduses a GRU network to learn user representationsfrom the sequence of clicked news.DKN (Wang et al., 2018), a knowledge-aware newsrecommendation method. It uses CNN to e content features used in our experiments are TF-IDFfeatures extracted from news texts.13news representations from news titles with bothword embeddings and entity embeddings (inferredfrom knowledge graph), and learns user representations based on the similarity between candidatenews and previously clicked news.NPA (Wu et al., 2019b), a neural news recommendation method with personalized attention mechanism to select important words and news articlesbased on user preferences to learn more informativenews and user representations.NAML (Wu et al., 2019a), a neural news recommendation method with attentive multi-view learning to incorporate different kinds of news information into the representations of news articles.LSTUR (An et al., 2019), a neural news recommendation method with long- and short-term userinterests. It models short-term user interest fromrecently clicked news with GRU and models longterm user interest from the whole click history.NRMS (Wu et al., 2019c), a neural news recommendation method which uses multi-head selfattention to learn news representations from thewords in news text and learn user representationsfrom previously clicked news articles.55.1ExperimentsExperimental SettingsIn our experiments, we verify and compare themethods introduced in Section 4 on the MINDdataset. Since most of these news recommendation methods are based on news titles, for fair comparison, we only used news titles in experimentsunless otherwise mentioned. We will explore theusefulness of different news texts such as body inSection 5.3.3. In order to simulate the practicalnews recommendation scenario where we alwayshave unseen users not included in training data, werandomly sampled half of the users for training, andused all the users for test. For those methods thatneed word embeddings, we used the Glove (Pennington et al., 2014) as initialization. Adam wasused as the optimizer. Since the non-clicked newsare usually much more than the clicked news ineach impression log, following (Wu et al., 2019b)we applied negative sampling technique to modeltraining. All hyper-parameters were selected according to the results on the validation set. Themetrics used in our experiments are AUC, MRR,nDCG@5 and nDCG@10, which are standard metrics for recommendation result evaluation. Eachexperiment was repeated 10 times.

URNRMSOverlap UsersUnseen 5240.6540.9441.3141.55Table 3: Results on the test set of the MIND dataset. Overlap users mean the users included in training set.5.2NAMLPerformance ComparisonThe experimental results of different methods onthe MIND dataset are summarized in Table 3. Wehave several findings from the results.First, news-specific recommendation methodssuch as NAML, LSTUR and NRMS usually performbetter than general recommendation methods likeWide&Deep, LibFM and DeepFM. This is becausein these news-specific recommendation methodsthe representations of news articles and user interest are learned in an end-to-end manner, while inthe general recommendation methods they are usually represented using handcrafted features. This result validates that learning representations of newscontent and user interest from raw data using neuralnetworks is more effective than feature engineering.The only exception is DFM, which is designed fornews recommendation but cannot outperform somegeneral recommendation methods such as DSSM.This is because in DFM the features of news andusers are also manually designed.Second, among the neural news recommendationmethods, NRMS can achieve the best performance.NRMS uses multi-head self-attention to capture therelatedness between words to learn news representations, and capture the interactions between previously clicked news articles to learn user representations. This result shows that advanced NLPmodels such as multi-head self-attention can effectively improve the understanding of news contentand modeling of user interest. The performanceof LSTUR is also strong. LSTUR can model users’short-term interest from their recently clicked newsthrough a GRU network, and model users’ longterm interest from the whole news click history.The result shows appropriate modeling of user interest is also important for news recommendation.Third, in terms of the AUC metric, the perfor-LDATF-IDFAvg-EmbAttentionCNNCNN AttSelf-Att.Self-Att AttLSTMLSTM 3136.4939.6639.1040.1040.0240.2340.2140.85Table 4: Different news representation methods. Attmeans attention mechanism.mance of news recommendation methods on unseen users is slightly lower than that on overlapusers which are included in training data. However,the performance on both kinds of users in terms ofMRR and nDCG metrics has no significant difference. This result indicates that by inferring userinterest from the content of their previously clickednews, the news recommendation models trainedon part of users can be effectively applied to theremaining users and new users coming in future.5.3News Content UnderstandingNext, we explore how to learn accurate news representations from textual content. Since the MINDdataset is quite large-scale, we randomly sampled100k samples from both training and test sets forthe experiments in this and the following sections.5.3.1News Representation ModelFirst, we compare different text representationmethods for learning news representation. We select three news recommendation methods whichhave strong performance, i.e., NAML, LSTUR andNRMS, and replace their original news representation module with different text representation methods, such as LDA, TF-IDF, average of word embed-

TitleAbs.BodyTitle Abs.Title Abs.Title Abs.Title Abs.Title Abs.Title Abs. Body (Con) Body Cat. (Con) Body Cat. Ent. (Con) Body (AMV) Body Cat. (AMV) Body Cat. Ent. 1.03Table 5: News representation with different news information. “Abs.”, “Cat.” and “Ent.” mean abstract,category and entity, respectively.Figure 3: BERT for news representation.ding (Avg-Emb), CNN, LSTM and multi-head selfattention (Self-Att). Since attention mechanism isan important technique in NLP (Yang et al., 2016),we also apply it to the aforementioned neural textrepresentation methods. The results are in Table 4.We have several findings from the results. First,neural text representation methods such as CNN,Self-Att and LSTM can outperform traditional textrepresentation methods like TF-IDF and LDA. Thisis because the neural text representation modelscan be learned with the news recommendation task,and they can capture the contexts of texts to generate better news representations. Second, Self-Attand LSTM outperform CNN in news representation. This is because multi-head self-attention andLSTM can capture long-range contexts of words,while CNN can only model local contexts. Third,the attention mechanism can effectively improvethe performance of different neural text representation methods such as CNN and LSTM for newsrecommendation. It shows that selecting important words in news texts using attention can helplearn more informative news representations. Another interesting finding is that the combination ofLSTM and attention can achieve the best performance. However, to our best knowledge, it is notused in existing news recommendation methods.5.3.2Pre-trained Language ModelsNext, we explore whether the quality of news representation can be further improved by the pretrained language models such as BERT (Devlinet al., 2019), which have achieved huge successin different NLP tasks. We applied BERT to thenews representation module of three state-of-theart news recommendation methods, i.e., NAML,LSTUR and NRMS. The results are summarizedin Fig. 3. We find that by replacing the origi-nal word embedding module with the pre-trainedBERT model, the performance of different newsrecommendation methods can be improved. Itshows the BERT model pre-trained on large-scalecorpus like Wikipedia can provide useful semanticinformation for news representation. We also findthat fine-tuning the pre-trained BERT model withthe news recommendation task can further improvethe performance. These results validate that thepre-trained language models are very helpful forunderstanding news articles.5.3.3 Different News InformationNext, we explore whether we can learn better newsrepresentation by incorporating more news information such as abstract and body. We try two methods for news text combination. The first one isdirect concatenation (denoted as Con), where wecombine different news texts into a long document.The second one is attentive multi-view learning(denoted as AMV) (Wu et al., 2019a) which modelseach kind of news text independently and combines them with an attention network. The resultsare shown in Table 5. We find that news bodiesare more effective than news titles and abstracts innews representation. This is because news bodiesare much longer and contain richer information ofnews content. Incorporating different kinds of newstexts such as title, body and abstract can effectivelyimprove the performance of news recommendation,indicating different news texts contain complementary information for news representation. Incorporating the category label and the entities in newstexts can further improve the performance. This isbecause category labels can provide general topicinformation, and the entities are keywords to understand the content of news. Another finding is thatthe attentive multi-view learning method is betterthan direct text combination in incorporating different news texts. This is because different news textsusually has different characteristics, and it is better

5.1239.3940.2439.8740.3340.3140.85Table 6: Different user modeling methods.to learn their representations using different neuralnetworks and model their different contributionsusing attention mechanisms.5.4User Interest ModelingMost of the state-of-the-art news recommendationmethods infer users’ interest in news from theirpreviously clicked news articles (Wu et al., 2019c;An et al.

3 MIND Dataset 3.1 Dataset Construction In order to facilitate the research in news recom-mendation, we built the MIcrosoft News Dataset (MIND)8. It was collected from the user behavior logs of Microsoft News9. We randomly sampled 1 million users who had at least 5 news click records during 6 weeks from October 12 to November 22, 2019.

Related Documents:

UCDP/PRIO Armed Conflict Dataset v.19.1, 1946-2018. UCDP Battle-Related Deaths Dataset v.19.1, 1989-2018. UCDP Dyadic Dataset v.19.1, 1946 - 2018. UCDP Non-State Conflict Dataset v.19.1, 1989-2018. UCDP One-sided Violence Dataset v.19.1, 1989-2018. UCDP Onset Dataset version 19.1, 1946-2018 UCDP Peace A

The Analysis Data Model Implementation Guide (ADaMIG) v1.1 defines three different types of datasets: analysis datasets, ADaM datasets, and non-ADaM analysis datasets: Analysis dataset - An analysis dataset is defined as a dataset used for analysis and reporting. ADaM dataset - An ADaM dataset is a particular type of analysis dataset that .

"Power BI Service Dataset" or Remote Dataset means dataset created in Power BI Service via REST API, sometimes referred as òOnline Dataset in Power BI community. Dataset includes in itself Power BI Tables and holds the configuration of relations between such tables. "Power BI Table" means a table inside Power BI Service Dataset.

CCC-466/SCALE 3 in 1985 CCC-725/SCALE 5 in 2004 CCC-545/SCALE 4.0 in 1990 CCC-732/SCALE 5.1 in 2006 SCALE 4.1 in 1992 CCC-750/SCALE 6.0 in 2009 SCALE 4.2 in 1994 CCC-785/SCALE 6.1 in 2011 SCALE 4.3 in 1995 CCC-834/SCALE 6.2 in 2016 The SCALE team is thankful for 40 years of sustaining support from NRC

The dataset uses the UCDP/PRIO Armed Conflict Dataset v.4-2011, 1946 – 2010 (Themnér & Wallensteen, 2011; also Gleditsch et al., 2002) as a base (and thus is an extension of the UCDP/PRIO dataset). Therefore, the dataset employs the UCDP/PRIO’s operational de

Trauma and Emergency Medical Services Information System Project (TEMSIS) The New Hampshire Bureau of Emergency Medical Services EMS Uniform PreHospital Dataset, Version 2.05 is composed of two separate documents or components. This primary document describes the EMS Dataset. A secondary EMS Demographic dataset has been defined which

The SketchTransfer training dataset consists of twoparts: labeled real images and unlabeled sketch images. Thetest dataset consists of labeled sketch images. To make thetask as straightforward as possible, we used the already pop-ular CIFAR-10 dataset as the source of labeled real images.We used the quickdraw dataset [23] as the source ofsketch .

Agile Software Development is not new, in fact it was introduced in the 1990s as a way to reduce costs, minimize risks and ensure that the final product is truly what customers requested. The idea behind the Agile approach is that instead of building a release that is huge in functionality (and often late to market), an organization would adapt to dynamic changing conditions by breaking a .