Pinned Repositories
BERT-for-Sequence-Labeling-and-Text-Classification
This is the template code to use BERT for sequence lableing and text classification, in order to facilitate BERT for more tasks. Currently, the template code has included conll-2003 named entity identification, Snips Slot Filling and Intent Prediction.
BERT_Paper_Chinese_Translation
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding 论文的中文翻译 Chinese Translation!
Entity-Relation-Extraction
Entity and Relation Extraction Based on TensorFlow and BERT. 基于TensorFlow和BERT的管道式实体及关系抽取,2019语言与智能技术竞赛信息抽取任务解决方案。Schema based Knowledge Extraction, SKE 2019
fan-ren-xiu-xian-zhuan
凡人修仙传(fanrenxiuxianzhuan)的资源汇总,谨献给“凡友”们。
Machine-Learning-Book
《机器学习宝典》包含:谷歌机器学习速成课程(招式)+机器学习术语表(口诀)+机器学习规则(心得)+机器学习中的常识性问题 (内功)。该资源适用于机器学习、深度学习研究人员和爱好者参考!
Machine_Learning_bookshelf
机器学习深度学习相关书籍、课件、代码的仓库。 Machine learning is the warehouse of books, courseware and codes.
Multimodal-short-video-dataset-and-baseline-classification-model
500,000 multimodal short video data and baseline models. 50万条多模态短视频数据集和基线模型(TensorFlow2.0)。
Multiple-Relations-Extraction-Only-Look-Once
Multiple-Relations-Extraction-Only-Look-Once. Just look at the sentence once and extract the multiple pairs of entities and their corresponding relations. 端到端联合多关系抽取模型,可用于 http://lic2019.ccf.org.cn/kg 信息抽取。
Schema-based-Knowledge-Extraction
Code for http://lic2019.ccf.org.cn/kg 信息抽取。使用基于 BERT 的实体抽取和关系抽取的端到端的联合模型。
Theoretical-Proof-of-Neural-Network-Model-and-Implementation-Based-on-Numpy
This resource implements a deep neural network through Numpy, and is equipped with easy-to-understand theoretical derivation, mainly for the in-depth understanding of neural networks. 神经网络模型的理论证明与基于Numpy的实现。
yuanxiaosc's Repositories
yuanxiaosc/SMP2018
SMP2018中文人机对话技术评测(ECDT)
yuanxiaosc/Keras_Attention_Seq2Seq
A sequence-to-sequence framework of Keras-based generative attention mechanisms that humans can read.一个人类可以阅读的基于Keras的代注意力机制的序列到序列的框架/模型,或许你不用写复杂的代码,直接使用吧。
yuanxiaosc/Deep_dynamic_contextualized_word_representation
TensorFlow code and pre-trained models for A Dynamic Word Representation Model Based on Deep Context. It combines the idea of BERT model and ELMo's deep context word representation.
yuanxiaosc/Seq2Seq-English-French-Machine-Translation-Model
Seq2Seq English-French Machine Translation Model
yuanxiaosc/TensorFlow_iris_tfrecord_and_estimator
This resource takes iris data set as an example to explain how to producte tfrecord files and use them in the customize Estimator!