- UID
- 15248
- 帖子
- 19457
- 积分
- 58820
- 学分
- 189332 个
- 金币
- 1000 个
- 性别
- 男
- 在线时间
- 20296 小时
|
2#
发表于 2017-9-3 23:14
| 只看该作者
There's also the introduction of a strange, pirate-like new character called Greenbeard.
书中还出现了一个名叫Greenbeard的怪咖,这个新角色的身份和海盗类似。
"It's obviously not perfect," Thoutt told Sam Hill over at Motherboard. "It isn't building a long-term story and the grammar isn't perfect. But the network is able to learn the basics of the English language and structure of George R.R. Martin's style on its own."
图特在接受Motherboard采访时告诉山姆•希尔,“这个算法显然并不完美,它不能编写长篇故事,语法也有问题。但是神经网络可以自学英语的基本语言知识以及马丁的文风结构。”
Neural networks are a type of machine learning algorithm that are inspired by the human brain's ability to not just memorize and follow instructions, but actually learn from past experiences.
神经网络是一种机器学习算法,设计灵感来自于人脑的记忆能力、遵循指令的能力以及从过去经验学习的能力。
A recurrent neural network is a specific subclass, which works best when it comes to processing long sequences of data, such as lengthy text from five previous books.
一个循环神经网络是一个特定的子集,最擅长处理长的数据序列,比如《冰与火之歌》前5部冗长的文本。
In theory, Thoutt's algorithm should be able to create a true sequel to Martin's existing work, based off things that have already happened in the novels.
理论上,图特的算法应该能基于书中已经出现的剧情创作出《冰与火之歌》真正的续集。
But in practice, the writing is clumsy and, most of the time, nonsensical. And it also references characters that have already died.
但实际上,这个算法的写作能力还很低级,大部分内容都不知所云,还会提到已经死掉的角色。
Still, some of the lines sound fairly prophetic:
不过,有些台词还是有一定预言性的:
"Arya saw Jon holding spears. Your grace," he said to an urgent maid, afraid. "The crow's eye would join you.
他对一个焦急的女仆说,“陛下,艾莉亚看到雪诺拿着长矛。乌鸦的眼睛会跟着你。”
"A perfect model would take everything that has happened in the books into account and not write about characters being alive when they died two books ago," Thoutt told Motherboard.
图特告诉Motherboard:“完美的算法模型能把书中的所有剧情考虑在内,且不会再让两部以前去世的角色再次复活。”
"The reality, though, is that the model isn't good enough to do that. If the model were that good authors might be in trouble ... but it makes a lot of mistakes because the technology to train a perfect text generator that can remember complex plots over millions of words doesn't exist yet."
“然而,实际上这个算法现在还不够完善。如果它有那么完美的话,作家们可能就要丢饭碗了……完美的文字创作机器可以记住数百万字的复杂剧情,现在的技术还不能训练出这种功能,它会犯很多错误。”
One of the main limitations here is the fact that the books just don't contain enough data for an algorithm.
最主要的局限之一是书中包含的数据对一个算法而言是不够的。
Although anyone who's read them will testify that they're pretty damn long, they actually represent quite a small data set for a neural network to learn from.
虽然《冰与火之歌》的读者都认为这部小说太长了,但是对于神经网络要学习的数据集来说,这些内容太少了。
But at the same time they contain a whole lot of unique words, nouns, and adjectives which aren't reused, which makes it very hard for the neural network to learn patterns.|
此外,书中包含了许多独特的词汇、名词和形容词,它们没有重复出现,这使得神经网络很难学习到模式。
Thoutt told Hill that a better source would be a book 100 times longer, but with the level of vocabulary of a children's book.
图特告诉希尔,更合适的数据源是一本比《冰与火之歌》长100倍,且词汇水平相当于儿童图书的书籍。
|
|