数据集地址:

  http://www.cs.cornell.edu/~cristian/Cornell_Movie-Dialogs_Corpus.html数据集介绍

  这个公开的资源被很多和自然语言处理NLP相关的开源代码和论文提到,

  所以仔细阅读了readme,并记录相关要点

  所有文件以" +++$+++ "分隔符

  - movie_titles_metadata.txt

  - 包含每部电影标题信息

  - fields:

  - movieID,

  - movie title,

  - movie year,

  - IMDB rating,

  - no. IMDB votes,

  - genres in the format ['genre1','genre2',?'genreN']

  - movie_characters_metadata.txt

  - 包含每部电影角色信息官网的telegram的的最新下载的地址

  - fields:

  - characterID

  - character name

  - movieID

  - movie title

  - gender ("?" for unlabeled cases)
telegram的下载网站在哪呢
  - position in credits ("?" for unlabeled cases)

  关键是下面两个文件,一个包含了所有文本,一个包含了文本之间的关系

  - movie_lines.txt

  - 包含每个表达(utterance)的实际文本

  - fields:

  - lineID

  - characterID (who uttered this phrase)

  - movieID

  - character name

  - text of the utterance

  前面5个样本:

  L1045 ++++++ u0 ++++++ m0 ++++++ BIANCA ++++++ They do not! L1044 ++++++ u2 ++++++ m0 ++++++ CAMERON ++++++ They do to! L985 ++++++ u0 ++++++ m0 ++++++ BIANCA ++++++ I hope so. L984 ++++++ u2 ++++++ m0 ++++++ CAMERON ++++++ She okay? L925 ++++++ u0 ++++++ m0 ++++++ BIANCA ++++++ Let's go.

  - movie_conversations.txt

  - 对话的结构-

  - fields

  - characterID of the first character involved in the conversation 对话中的第一个角色的ID

  - characterID of the second character involved in the conversation 对话中的第二个角色的ID

  - movieID of the movie in which the conversation occurred 对话所属电影的ID

  - list of the utterances that make the conversation, in chronological

  order: ['lineID1','lineID2',?'lineIDN']

  has to be matched with movie_lines.txt to reconstruct the actual content

  对话中以时间顺序的各个表达的列表,

  order: ['lineID1','lineID2',?'lineIDN']必须和movie_lines.txt匹配以便于重构实际内容

  前面5个样本:

  u0 ++++++ u2 ++++++ m0 ++++++ ['L194', 'L195', 'L196', 'L197'] u0 ++++++ u2 ++++++ m0 ++++++ ['L198', 'L199'] u0 ++++++ u2 ++++++ m0 ++++++ ['L200', 'L201', 'L202', 'L203'] u0 ++++++ u2 ++++++ m0 ++++++ ['L204', 'L205', 'L206'] u0 ++++++ u2 ++++++ m0 +++

  - raw_script_urls.txt

  -原始来源的url( the urls from which the raw sources were retrieved)

  ========================================================================================英文版:

  Cornell Movie-Dialogs Corpus

  Distributed together with:

  "Chameleons in imagined conversations: A new approach to understanding coordination of linguistic style in dialogs"

  Cristian Danescu-Niculescu-Mizil and Lillian Lee

  Proceedings of 无障碍的telegram的下载网址在哪呢 the Workshop on Cognitive Modeling and Computational Linguistics, ACL 2011.

  (this paper is included in this zip file)

  NOTE: If you have results to report on these corpora, please send email to cristian@cs.cornell.edu or llee@cs.cornell.edu so we can add you to our list of people using this data. Thanks!

  Contents of this README:

  A) Brief description

  B) Files description

  C) Details on the collection procedure

  D) Contact

  A) Brief description:

  This corpus contains a metadata-rich collection of fictional conversations extracted from raw movie scripts:

  - 220,579 conversational exchanges between 10,292 pairs of movie characters

  - involves 9,035 characters from 617 movies

  - in total 304,713 utterances

  - movie metadata included:

  - genres

  - release year

  - IMDB rating

  - number of IMDB votes

  - IMDB rating

  - character metadata included:

  - gender (for 3,774 characters)

  - position on movie credits (3,321 characters)

  B) Files description:

  In all files the field separator is " +++$+++ "

  - movie_titles_metadata.txt

  - contains information about each movie title

  - fields:

  - movieID,

  - movie title,

  - movie year,

  - IMDB rating,

  - no. IMDB votes,

  - genres in the format ['genre1','genre2',É,'genreN']

  - movie_characters_metadata.txt

  - contains information about each movie character

  - fields:

  - characterID

  - character name

  - movieID

  - movie title

  - gender ("?" for unlabeled cases)

  - position in credits ("?" for unlabeled cases)

  - movie_lines.txt

  - 官方的最新版telegram的下载入口是多少 contains the actual text of each utterance

  - fields:

  - lineID

  - characterID (who uttered this phrase)

  - movieID

  - character name

  - text of the utterance

  - movie_conversations.txt

  - the structure of the conversations

  - fields

  - characterID of the first character involved in the conversation

  - characterID of 最新的官网telegram下载地方哪里有 the second character involved in the conversation
官方telegram 网站下载地方在哪里
  - movieID of the movie in which the conversation occurred

  - list of the utterances that make the conversation, in chronological

  order: 最新的中文版的telegram的下载入口在哪呢 ['lineID1','lineID2',É,'lineIDN']

  has to be matched with movie_lines.txt to reconstruct the actual content

  - raw_script_urls.txt

  - the urls from which the raw sources were retrieved

  C) Details on the collection procedure:

  We started from raw publicly available movie scripts (sources acknowledged in

  raw_script_urls.txt). In order to collect the metadata necessary for this study

  and to distinguish between two script versions of the same movie, we automatically

  matched each script with an entry in movie database provided by IMDB (The Internet

  Movie Database; data interfaces available at http://www.imdb.com/interfaces). Some

  amount of manual correction was also involved. When more than one movie with the same

  title was found in IMBD, the match was made with the most popular title

  (the one that received most IMDB votes)

  After discarding all movies that could not be matched or that had less than 5 IMDB

  votes, we were left with 617 unique titles with metadata including genre, release

  year, IMDB rating and no. of IMDB votes and cast distribution. We then identified

  the pairs of characters that interact and separated their conversations automatically

  using simple data processing heuristics. After discarding all pairs that exchanged

  less than 5 conversational exchanges there were 10,292 left, exchanging 220,579

  conversational exchanges (304,713 utterances). After automatically matching the names

  of the 9,035 involved characters to the list of cast distribution, we used the

  gender of each interpreting actor to infer the fictional gender of a subset of

  3,321 movie characters (we raised the number of gendered 3,774 characters through

  manual annotation). Similarly, we collected the end credit position of a subset

  of 3,321 characters as a proxy for their status.

  D) Contact:

  Please email any questions to: cristian@cs.cornell.edu 无障碍telegram下载网址是什么 (Cristian Danescu-Niculescu-Mizil)

seo