Git Product home page Git Product logo

Comments (12)

honnibal avatar honnibal commented on August 10, 2024

The problem is that the requirements.txt file hasn't version pinned for the dependencies. I've updated/rewritten these libraries, and now redshift doesn't compile.

The offending change seems to be this one:

1487f64

Instead of updating the pin to the correct version, I removed the version pin. Damn.

To solve this, the correct versions need to be identified, and the versions specified in the requirements.txt

The following libraries need version pinning:

murmurhash
cymem
preshed
thinc

I've made lots of releases of thinc, and a couple of preshed. cymem and murmurhash have been more stable.

Looking through the commit history of thinc, and matching up the dates, it looks like v1.73 is a likely candidate.

from redshift.

honnibal avatar honnibal commented on August 10, 2024

Found the right versions. Now compiles for me. Give it a go.

Btw, just checking that you know that this library is code from my research, and that the maintained library is spaCy? Probably this library is only useful for replicating or extending one of my papers, especially the disfluency detection one. If you're doing that, let me know if you have any questions.

But if you're just looking for a good parsing library, try spaCy :). http://spacy.io

from redshift.

pawanjay176 avatar pawanjay176 commented on August 10, 2024

I tried installing it again. Gives me the following errors now
http://pastebin.com/7T7tV0xN

I just wanted a fast POS tagging utility. I was not aware of spaCy. Using that now. Thanks a lot for the quick reply :)

from redshift.

corei5 avatar corei5 commented on August 10, 2024

http://nlp.stanford.edu/software/tagger.shtml

On Fri, Dec 4, 2015 at 9:53 PM, pawanjay176 [email protected]
wrote:

I tried installing it again. Gives me the following errors now
http://pastebin.com/7T7tV0xN

I just wanted a fast POS tagging utility. I was not aware of spaCy. Using
that now. Thanks a lot for the quick reply :)


Reply to this email directly or view it on GitHub
#20 (comment).

from redshift.

honnibal avatar honnibal commented on August 10, 2024

spaCy's tagger is MIT licensed, can be used from Python, and is much faster than Stanford's :). Accuracy should be the same as well.

from redshift.

corei5 avatar corei5 commented on August 10, 2024

how can i use spaCy's or stanford for for parsing only for define noun and
verb ?.i want to send noun and verb in a database because i want to use it
search poses in a search engine .

On Fri, Dec 4, 2015 at 10:06 PM, Matthew Honnibal [email protected]
wrote:

spaCy's tagger can be used from Python, and is much faster than Stanford's
:). Accuracy should be the same as well.


Reply to this email directly or view it on GitHub
#20 (comment).

from redshift.

honnibal avatar honnibal commented on August 10, 2024

I'm not sure I understand your question. But this will load spaCy's default English model, analyse some text, and return only the nouns and verbs. The parser is here disabled for efficiency.

>>> from spacy.en import English
>>> from spacy.attrs import NOUN, VERB
>>> nlp = English(parser=False)
>>> doc = nlp(u'An example sentence, that has two nouns. This sentence contains three nouns.')
>>> print([w.text for w in doc if w.pos in (NOUN, VERB)]
[u'sentence', u'has', u'nouns', u'contains', u'nouns']

See the spaCy docs for details. You should ask further questions there as well :)

from redshift.

corei5 avatar corei5 commented on August 10, 2024

great....i want to make a search engine for my BSC final year project and
it is ontology base semantic search engine . it depends on verb . the main
class is verb for search and sub class in noun .so i need to define verb
and noun and search in a specific domain .
do you understand ,about my project . i really need parsing it for future
work .

On Fri, Dec 4, 2015 at 10:59 PM, Matthew Honnibal [email protected]
wrote:

I'm not sure I understand your question. But this will load spaCy's
default English model, analyse some text, and return only the nouns and
verbs. The parser is here disabled for efficiency.

from spacy.en import English
from spacy.attrs import NOUN, VERB
nlp = English(parser=False)
doc = nlp(u'An example sentence, that has two nouns. This sentence contains three nouns.')
print([w.text for w in doc if w.pos in (NOUN, VERB)]
[u'sentence', u'has', u'nouns', u'contains', u'nouns']

See the spaCy docs for details. You should ask further questions there as
well :)


Reply to this email directly or view it on GitHub
#20 (comment).

from redshift.

pawanjay176 avatar pawanjay176 commented on August 10, 2024

Thanks a ton!
Tagger is really accurate and way way fast than stanford tagger

from redshift.

corei5 avatar corei5 commented on August 10, 2024

but i do not understand how can i use it for search engine?can you kindly
describe it.plz

On Sat, Dec 5, 2015 at 1:03 AM, pawanjay176 [email protected]
wrote:

Thanks a ton!
Tagger is really accurate and way way fast than stanford tagger


Reply to this email directly or view it on GitHub
#20 (comment).

from redshift.

pawanjay176 avatar pawanjay176 commented on August 10, 2024

You can pos_tag any sentence like this

from spacy.en import English, LOCAL_DATA_DIR
import spacy.en
import os, time
data_dir = os.environ.get('SPACY_DATA', LOCAL_DATA_DIR)
nlp = English(parser=False, tagger=True, entity=False)

def print_fine_pos(token):
return (token.tag_)

def pos_tags(sentence):
sentence = unicode(sentence, "utf-8")
tokens = nlp(sentence)
tags = []
for tok in tokens:
tags.append((tok,print_fine_pos(tok)))
return tags

print pos_tags("This is a sentence")

from redshift.

corei5 avatar corei5 commented on August 10, 2024

how can i install spaCy.io in Ubuntu 14.04 ? i install Canopy before
.

On Sat, Dec 5, 2015 at 1:25 AM, pawanjay176 [email protected]
wrote:

You can pos_tag any sentence like this

from spacy.en import English, LOCAL_DATA_DIR
import spacy.en
import os, time
data_dir = os.environ.get('SPACY_DATA', LOCAL_DATA_DIR)
nlp = English(parser=False, tagger=True, entity=False)

def print_fine_pos(token):
return (token.tag_)

def pos_tags(sentence):
sentence = unicode(sentence, "utf-8")
tokens = nlp(sentence)
tags = []
for tok in tokens:
tags.append((tok,print_fine_pos(tok)))
return tags

print pos_tags("This is a sentence")


Reply to this email directly or view it on GitHub
#20 (comment).

from redshift.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.