Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Tokenization of URLs needs work #344

Closed
rlvoyer opened this issue Apr 19, 2016 · 2 comments · Fixed by #700
Closed

Tokenization of URLs needs work #344

rlvoyer opened this issue Apr 19, 2016 · 2 comments · Fixed by #700

Comments

@rlvoyer
Copy link

rlvoyer commented Apr 19, 2016

In [3]: nlp = English()

In [4]: doc = nlp("Do you agree that this is a URL: http://www.nytimes.com/2016/04/20/us/politics/new-york-primary-preview.html?hp&action=click&pgtype=Homepage&clickSource=story-heading&module=a-lede-package-region&region=top-news&WT.nav=top-news&_r=0")

In [5]: [s.lemma_.lower() for s in doc if not s.like_url]
Out[5]:
['do',
 'you',
 'agree',
 'that',
 'this',
 'be',
 'a',
 'url',
 ':',
 '-',
 'york-primary-preview.html?hp&action=click&pgtype=homepage&clicksource=story-heading&module=a-lede-package-region&region=top-news&wt.nav=top-news&_r=0']
@rlvoyer
Copy link
Author

rlvoyer commented Apr 19, 2016

In fact, the problem here might be better characterized as a tokenization problem (so I'm going to rename the issue).

@rlvoyer rlvoyer changed the title URL identification needs work Tokenization of URLs needs work Apr 19, 2016
@lock
Copy link

lock bot commented May 9, 2018

This thread has been automatically locked since there has not been any recent activity after it was closed. Please open a new issue for related bugs.

@lock lock bot locked as resolved and limited conversation to collaborators May 9, 2018
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants