Merge Zixaphir X

This commit is contained in:
seaweedchan 2013-08-13 05:28:46 -07:00
commit f2b15cd0ab
4 changed files with 44 additions and 40 deletions

View File

@ -142,7 +142,6 @@
},
'Linkification': {
'Linkify': [true, 'Convert text into links where applicable.'],
'Allow False Positives': [false, 'Linkify everything, allowing more false positives but reducing missed links'],
'Embedding': [true, 'Embed supported services.'],
'Auto-embed': [false, 'Auto-embed Linkify Embeds.'],
'Link Title': [true, 'Replace the link of a supported site with its actual title. Currently Supported: YouTube, Vimeo, SoundCloud, and Github gists']
@ -4268,7 +4267,7 @@
if (g.VIEW === 'catalog' || !Conf['Linkify']) {
return;
}
this.regString = Conf['Allow False Positives'] ? /([-a-z]+:\/\/|[a-z]{3,}\.[-a-z0-9]+\.[a-z]|[-a-z0-9]+\.[a-z]|[\d]+\.[\d]+\.[\d]+\.[\d]+\/|[a-z]{3,}:[a-z0-9?]|[^\s@]+@[a-z0-9.-]+\.[a-z0-9])/i : /(((magnet|mailto)\:|(www\.)|(news|(ht|f)tp(s?))\:\/\/){1})/i;
this.regString = /(?:[a-z][-\w]+:([a-z\d%\/])|www\d{0,3}[.]|[-a-z\d.]+[.](com|net|org|jp|uk|ru|be|tv|xxx|edu|gov|cd|es|de|se|tk|dk|io|fm|fi)|[\d]{1,3}\.[\d]{1,3}\.[\d]{1,3}\.[\d]{1,3}|[-\w\d.@]+@[a-z\d.-]+\.[a-z\d])/i;
if (Conf['Comment Expansion']) {
ExpandComment.callbacks.push(this.node);
}
@ -4299,8 +4298,8 @@
space = /[\s'"]/;
snapshot = $.X('.//br|.//text()', this.nodes.comment);
i = 0;
links = [];
while (node = snapshot.snapshotItem(i++)) {
links = [];
data = node.data;
if (node.parentElement.nodeName === "A" || !data) {
continue;
@ -4316,7 +4315,7 @@
endNode = saved;
length = saved.data.length;
if (end = space.exec(saved.data)) {
length = end.index;
test.lastIndex = length = end.index;
i--;
break;
}
@ -4328,22 +4327,23 @@
if (link = Linkify.regString.exec(text = range.toString())) {
if (lIndex = link.index) {
range.setStart(node, lIndex + index);
text = text.slice(0, lIndex);
}
links.push([range, text]);
}
break;
} else {
if (link = Linkify.regString.exec(result[0])) {
range = Linkify.makeRange(node, node, link.index, link.length);
range = Linkify.makeRange(node, node, index + link.index, length + link.index);
links.push([range, link]);
}
}
}
_ref = links.reverse();
for (_i = 0, _len = _ref.length; _i < _len; _i++) {
range = _ref[_i];
this.nodes.links.push(Linkify.makeLink(range, this));
}
}
_ref = links.reverse();
for (_i = 0, _len = _ref.length; _i < _len; _i++) {
range = _ref[_i];
this.nodes.links.push(Linkify.makeLink(range, this));
}
if (!(Conf['Embedding'] || Conf['Link Title'])) {
return;

View File

@ -124,7 +124,6 @@
},
'Linkification': {
'Linkify': [true, 'Convert text into links where applicable.'],
'Allow False Positives': [false, 'Linkify everything, allowing more false positives but reducing missed links'],
'Embedding': [true, 'Embed supported services.'],
'Auto-embed': [false, 'Auto-embed Linkify Embeds.'],
'Link Title': [true, 'Replace the link of a supported site with its actual title. Currently Supported: YouTube, Vimeo, SoundCloud, and Github gists']
@ -4274,7 +4273,7 @@
if (g.VIEW === 'catalog' || !Conf['Linkify']) {
return;
}
this.regString = Conf['Allow False Positives'] ? /([-a-z]+:\/\/|[a-z]{3,}\.[-a-z0-9]+\.[a-z]|[-a-z0-9]+\.[a-z]|[\d]+\.[\d]+\.[\d]+\.[\d]+\/|[a-z]{3,}:[a-z0-9?]|[^\s@]+@[a-z0-9.-]+\.[a-z0-9])/i : /(((magnet|mailto)\:|(www\.)|(news|(ht|f)tp(s?))\:\/\/){1})/i;
this.regString = /(?:[a-z][-\w]+:([a-z\d%\/])|www\d{0,3}[.]|[-a-z\d.]+[.](com|net|org|jp|uk|ru|be|tv|xxx|edu|gov|cd|es|de|se|tk|dk|io|fm|fi)|[\d]{1,3}\.[\d]{1,3}\.[\d]{1,3}\.[\d]{1,3}|[-\w\d.@]+@[a-z\d.-]+\.[a-z\d])/i;
if (Conf['Comment Expansion']) {
ExpandComment.callbacks.push(this.node);
}
@ -4305,8 +4304,8 @@
space = /[\s'"]/;
snapshot = $.X('.//br|.//text()', this.nodes.comment);
i = 0;
links = [];
while (node = snapshot.snapshotItem(i++)) {
links = [];
data = node.data;
if (node.parentElement.nodeName === "A" || !data) {
continue;
@ -4322,7 +4321,7 @@
endNode = saved;
length = saved.data.length;
if (end = space.exec(saved.data)) {
length = end.index;
test.lastIndex = length = end.index;
i--;
break;
}
@ -4334,22 +4333,23 @@
if (link = Linkify.regString.exec(text = range.toString())) {
if (lIndex = link.index) {
range.setStart(node, lIndex + index);
text = text.slice(0, lIndex);
}
links.push([range, text]);
}
break;
} else {
if (link = Linkify.regString.exec(result[0])) {
range = Linkify.makeRange(node, node, link.index, link.length);
range = Linkify.makeRange(node, node, index + link.index, length + link.index);
links.push([range, link]);
}
}
}
_ref = links.reverse();
for (_i = 0, _len = _ref.length; _i < _len; _i++) {
range = _ref[_i];
this.nodes.links.push(Linkify.makeLink(range, this));
}
}
_ref = links.reverse();
for (_i = 0, _len = _ref.length; _i < _len; _i++) {
range = _ref[_i];
this.nodes.links.push(Linkify.makeLink(range, this));
}
if (!(Conf['Embedding'] || Conf['Link Title'])) {
return;

View File

@ -83,10 +83,6 @@ Config =
true
'Convert text into links where applicable.'
]
'Allow False Positives': [
false
'Linkify everything, allowing more false positives but reducing missed links'
]
'Embedding': [
true
'Embed supported services.'

View File

@ -2,22 +2,27 @@ Linkify =
init: ->
return if g.VIEW is 'catalog' or not Conf['Linkify']
@regString = if Conf['Allow False Positives']
@regString =
///(
[-a-z]+://
# http, magnet, ftp, etc
?:[a-z][-\w]+:(
[a-z\d%/]
)
|
[a-z]{3,}\.[-a-z0-9]+\.[a-z]
www\d{0,3}[.]
|
[-a-z0-9]+\.[a-z]
# This should account for virtually all links posted without www or http:
# If it misses any, screw it. No, I will not add canv.as
[-a-z\d.]+[.](
com|net|org|jp|uk|ru|be|tv|xxx|edu|gov|cd|es|de|se|tk|dk|io|fm|fi
)
|
[\d]+\.[\d]+\.[\d]+\.[\d]+/
# IPv4 Addresses
[\d]{1,3}\.[\d]{1,3}\.[\d]{1,3}\.[\d]{1,3}
|
[a-z]{3,}:[a-z0-9?]
|
[^\s@]+@[a-z0-9.-]+\.[a-z0-9]
# E-mails
[-\w\d.@]+@[a-z\d.-]+\.[a-z\d]
)///i
else
/(((magnet|mailto)\:|(www\.)|(news|(ht|f)tp(s?))\:\/\/){1})/i
if Conf['Comment Expansion']
ExpandComment.callbacks.push @node
@ -45,14 +50,15 @@ Linkify =
snapshot = $.X './/br|.//text()', @nodes.comment
i = 0
links = []
while node = snapshot.snapshotItem i++
links = []
{data} = node
continue if node.parentElement.nodeName is "A" or not data
while result = test.exec data
{index} = result
endNode = node
# End of node, not necessarily end of space-delimited string
if (length = index + result[0].length) is data.length
while (saved = snapshot.snapshotItem i++)
@ -62,25 +68,27 @@ Linkify =
{length} = saved.data
if end = space.exec saved.data
length = end.index
# Set our snapshot and regex to start on this node at this position when the loop resumes
test.lastIndex = length = end.index
i--
break
if length is endNode.data.length then test.lastIndex = 0
test.lastIndex = 0 if length is endNode.data.length
range = Linkify.makeRange node, endNode, index, length
if link = Linkify.regString.exec text = range.toString()
if lIndex = link.index
range.setStart node, lIndex + index
text = text[...lIndex]
links.push [range, text]
break
else
if link = Linkify.regString.exec result[0]
range = Linkify.makeRange node, node, link.index, link.length
range = Linkify.makeRange node, node, index + link.index, length + link.index
links.push [range, link]
for range in links.reverse()
@nodes.links.push Linkify.makeLink range, @
for range in links.reverse()
@nodes.links.push Linkify.makeLink range, @
return unless Conf['Embedding'] or Conf['Link Title']