您好,欢迎光临本网站![请登录][注册会员]  

搜索资源列表

  1. Lucene+in+Action简体中文版1-4章

  2. 4.1 使用分析器 分析(Analysis),在Lucene当中指的是将域(Field)文本转换为最基本的索引表示单元——项(Term)的过程。在搜索过程中,这些项用于决定什么样的文档能匹配查询条件。例如,如果这句话“For example, if this sentence were indexed into a field”被索引到一个域(Field)中(假设域类型为Field.Text),相应产生的项可能是以for和example两个单词打头,其它的项随之按照它在句子中出现的先后顺序逐个
  3. 所属分类:专业指导

    • 发布日期:2009-08-13
    • 文件大小:2097152
    • 提供者:turing2015
  1. SPEECH and LANGUAGE PROCESSING

  2. SPEECH and LANGUAGE PROCESSING An Introduction to Natural Language Processing, Computational Linguistics, and Speech Recognition Second Edition by Daniel Jurafsky and James H. Martin Last Update January 6, 2009 The 2nd edition is now avaiable. A mil
  3. 所属分类:C++

    • 发布日期:2010-07-05
    • 文件大小:14680064
    • 提供者:wx90wx
  1. StringToken

  2. String tokenization is defined as the problem that consists of breaking up a string into tokens which are seperated by delimiters. Both tokens and delimiters are themselves strings. Commonly used string structures that require the use of string toke
  3. 所属分类:其它

    • 发布日期:2006-02-23
    • 文件大小:4194304
    • 提供者:chenxh
  1. English tokenization

  2. Moses中自带的英文分词,大家可以下载使用处理英文的分词,具体的使用命令就是tokenizer.perl example.txt
  3. 所属分类:Perl

    • 发布日期:2012-02-06
    • 文件大小:19456
    • 提供者:tianliang0123
  1. Apache Solr 3 Enterprise Search Server

  2. Design a schema to include text indexing details like tokenization, stemming, and synonyms Import data using various formats like CSV, XML, and from databases, and extract text from common document formats Search using Solr’s rich query syntax, perf
  3. 所属分类:Web服务器

    • 发布日期:2012-04-28
    • 文件大小:8388608
    • 提供者:dongweizu33
  1. Hibernate Search in Sction

  2. 1 State of the art 3 1.1 What is search? 4 Categorizing information 5 ■ Using a detailed search screen 5 Using a user-friendly search box 7 ■ Mixing search strategies 7 Choosing a strategy: the first step on a long road 8 1.2 Pitfalls of search engi
  3. 所属分类:Java

    • 发布日期:2013-03-25
    • 文件大小:8388608
    • 提供者:quebu1989
  1. The Lancaster Corpus of Mandarin Chinese

  2. The Lancaster Corpus of Mandarin Chinese (LCMC) is designed as a Chinese match for the FLOB and FROWN corpora for modern British and American English. The corpus is suitable for use in both monolingual research into modern Mandarin Chinese and cross
  3. 所属分类:其它

    • 发布日期:2013-04-01
    • 文件大小:5242880
    • 提供者:u010128875
  1. stanford chinese segmentor

  2. Tokenization of raw text is a standard pre-processing step for many NLP tasks. For English, tokenization usually involves punctuation splitting and separation of some affixes like possessives. Other languages require more extensive token pre-process
  3. 所属分类:企业管理

    • 发布日期:2013-04-10
    • 文件大小:1048576
    • 提供者:bbking1981
  1. Tokenization 分析及研究

  2. Token技术产生的背景、解决的问题及应用分析
  3. 所属分类:金融

    • 发布日期:2014-11-05
    • 文件大小:368640
    • 提供者:alfredmeng
  1. Python 3 Text Processing with NLTK 3 Cookbook

  2. http://www.amazon.com/Python-Text-Processing-NLTK-Cookbook/dp/1782167854/ Paperback: 310 pages Publisher: Packt Publishing - ebooks Account (August 26, 2014) Language: English Over 80 practical recipes on natural language processing techniques using
  3. 所属分类:Python

    • 发布日期:2015-01-21
    • 文件大小:1048576
    • 提供者:u011433684
  1. Python 3 Text Processing with NLTK 3 Cookbook(PACKT,2014)

  2. This book will show you the essential techniques of text and language processing. Starting with tokenization, stemming, and the WordNet dictionary, you'll progress to part-of-speech tagging, phrase chunking, and named entity recognition. You'll lear
  3. 所属分类:Python

    • 发布日期:2015-04-17
    • 文件大小:1048576
    • 提供者:vanridin
  1. 基于python的NLTK

  2. nltk3.0 NLTK is a leading platform for building Python programs to work with human language data. It provides easy-to-use interfaces to over 50 corpora and lexical resources such as WordNet, along with a suite of text processing libraries for classi
  3. 所属分类:Python

    • 发布日期:2015-04-23
    • 文件大小:1048576
    • 提供者:u011562970
  1. NLTK.Essentials

  2. Build cool NLP and machine learning applications using NLTK and other Python libraries About This Book Extract information from unstructured data using NLTK to solve NLP problems Analyse linguistic structures in text and learn the concept of semanti
  3. 所属分类:互联网

    • 发布日期:2015-09-21
    • 文件大小:2097152
    • 提供者:ramissue
  1. NLTP 3 python 3

  2. This book will show you the essential techniques of text and language processing. Starting with tokenization, stemming, and the WordNet dictionary, you'll progress to part-of-speech tagging, phrase chunking, and named entity recognition. You'll lear
  3. 所属分类:其它

    • 发布日期:2016-01-16
    • 文件大小:142336
    • 提供者:junjian_1212
  1. Artificial Intelligence With Python[January 2017]

  2. Artificial intelligence is becoming increasingly relevant in the modern world where everything is driven by data and automation. It is used extensively across many fields such as image recognition, robotics, search engines, and self-driving cars. In
  3. 所属分类:Python

    • 发布日期:2017-11-25
    • 文件大小:33554432
    • 提供者:fantasy_alex
  1. Critical Tokenization and its Properties

  2. 郭进guo jin博士论文,关于自然语言处理
  3. 所属分类:专业指导

    • 发布日期:2009-03-01
    • 文件大小:1048576
    • 提供者:chengsshi
  1. 自然语言处理和机器翻译手册

  2. Contents 1 Data Acquisition and Linguistic Resources ............................................................... 1 1.1 Introduction ............................................................................................................1 1.2
  3. 所属分类:机器学习

    • 发布日期:2018-10-04
    • 文件大小:31457280
    • 提供者:u010652219
  1. cpp-YouTokenToMe高性能无监督文本标记化tokenization工具

  2. YouTokenToMe:高性能无监督文本标记化(tokenization)工具
  3. 所属分类:其它

    • 发布日期:2019-08-16
    • 文件大小:84992
    • 提供者:weixin_39841848
  1. tokenization-源码

  2. tokenization
  3. 所属分类:其它

    • 发布日期:2021-03-11
    • 文件大小:494592
    • 提供者:weixin_42128988
  1. TOKENIZATION-源码

  2. TOKENIZATION
  3. 所属分类:其它

    • 发布日期:2021-03-26
    • 文件大小:4096
    • 提供者:weixin_42109732
« 12 »