|
|
|
@ -10,6 +10,7 @@ import ( |
|
|
|
|
"path" |
|
|
|
|
"path/filepath" |
|
|
|
|
"regexp" |
|
|
|
|
"slices" |
|
|
|
|
"strings" |
|
|
|
|
"sync" |
|
|
|
|
|
|
|
|
@ -54,7 +55,7 @@ var ( |
|
|
|
|
shortLinkPattern = regexp.MustCompile(`\[\[(.*?)\]\](\w*)`) |
|
|
|
|
|
|
|
|
|
// anyHashPattern splits url containing SHA into parts
|
|
|
|
|
anyHashPattern = regexp.MustCompile(`https?://(?:\S+/){4,5}([0-9a-f]{40,64})(/[-+~_%.a-zA-Z0-9/]+)?(#[-+~_%.a-zA-Z0-9]+)?`) |
|
|
|
|
anyHashPattern = regexp.MustCompile(`https?://(?:\S+/){4,5}([0-9a-f]{40,64})(/[-+~%./\w]+)?(\?[-+~%.\w&=]+)?(#[-+~%.\w]+)?`) |
|
|
|
|
|
|
|
|
|
// comparePattern matches "http://domain/org/repo/compare/COMMIT1...COMMIT2#hash"
|
|
|
|
|
comparePattern = regexp.MustCompile(`https?://(?:\S+/){4,5}([0-9a-f]{7,64})(\.\.\.?)([0-9a-f]{7,64})?(#[-+~_%.a-zA-Z0-9]+)?`) |
|
|
|
@ -591,7 +592,8 @@ func replaceContentList(node *html.Node, i, j int, newNodes []*html.Node) { |
|
|
|
|
|
|
|
|
|
func mentionProcessor(ctx *RenderContext, node *html.Node) { |
|
|
|
|
start := 0 |
|
|
|
|
for node != nil { |
|
|
|
|
nodeStop := node.NextSibling |
|
|
|
|
for node != nodeStop { |
|
|
|
|
found, loc := references.FindFirstMentionBytes(util.UnsafeStringToBytes(node.Data[start:])) |
|
|
|
|
if !found { |
|
|
|
|
node = node.NextSibling |
|
|
|
@ -962,57 +964,68 @@ func commitCrossReferencePatternProcessor(ctx *RenderContext, node *html.Node) { |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
// fullHashPatternProcessor renders SHA containing URLs
|
|
|
|
|
func fullHashPatternProcessor(ctx *RenderContext, node *html.Node) { |
|
|
|
|
if ctx.Metas == nil { |
|
|
|
|
return |
|
|
|
|
type anyHashPatternResult struct { |
|
|
|
|
PosStart int |
|
|
|
|
PosEnd int |
|
|
|
|
FullURL string |
|
|
|
|
CommitID string |
|
|
|
|
SubPath string |
|
|
|
|
QueryHash string |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
func anyHashPatternExtract(s string) (ret anyHashPatternResult, ok bool) { |
|
|
|
|
m := anyHashPattern.FindStringSubmatchIndex(s) |
|
|
|
|
if m == nil { |
|
|
|
|
return ret, false |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
next := node.NextSibling |
|
|
|
|
for node != nil && node != next { |
|
|
|
|
m := anyHashPattern.FindStringSubmatchIndex(node.Data) |
|
|
|
|
if m == nil { |
|
|
|
|
return |
|
|
|
|
ret.PosStart, ret.PosEnd = m[0], m[1] |
|
|
|
|
ret.FullURL = s[ret.PosStart:ret.PosEnd] |
|
|
|
|
if strings.HasSuffix(ret.FullURL, ".") { |
|
|
|
|
// if url ends in '.', it's very likely that it is not part of the actual url but used to finish a sentence.
|
|
|
|
|
ret.PosEnd-- |
|
|
|
|
ret.FullURL = ret.FullURL[:len(ret.FullURL)-1] |
|
|
|
|
for i := 0; i < len(m); i++ { |
|
|
|
|
m[i] = min(m[i], ret.PosEnd) |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
urlFull := node.Data[m[0]:m[1]] |
|
|
|
|
text := base.ShortSha(node.Data[m[2]:m[3]]) |
|
|
|
|
ret.CommitID = s[m[2]:m[3]] |
|
|
|
|
if m[5] > 0 { |
|
|
|
|
ret.SubPath = s[m[4]:m[5]] |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
// 3rd capture group matches a optional path
|
|
|
|
|
subpath := "" |
|
|
|
|
if m[5] > 0 { |
|
|
|
|
subpath = node.Data[m[4]:m[5]] |
|
|
|
|
} |
|
|
|
|
lastStart, lastEnd := m[len(m)-2], m[len(m)-1] |
|
|
|
|
if lastEnd > 0 { |
|
|
|
|
ret.QueryHash = s[lastStart:lastEnd][1:] |
|
|
|
|
} |
|
|
|
|
return ret, true |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
// 4th capture group matches a optional url hash
|
|
|
|
|
hash := "" |
|
|
|
|
if m[7] > 0 { |
|
|
|
|
hash = node.Data[m[6]:m[7]][1:] |
|
|
|
|
// fullHashPatternProcessor renders SHA containing URLs
|
|
|
|
|
func fullHashPatternProcessor(ctx *RenderContext, node *html.Node) { |
|
|
|
|
if ctx.Metas == nil { |
|
|
|
|
return |
|
|
|
|
} |
|
|
|
|
nodeStop := node.NextSibling |
|
|
|
|
for node != nodeStop { |
|
|
|
|
if node.Type != html.TextNode { |
|
|
|
|
node = node.NextSibling |
|
|
|
|
continue |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
start := m[0] |
|
|
|
|
end := m[1] |
|
|
|
|
|
|
|
|
|
// If url ends in '.', it's very likely that it is not part of the
|
|
|
|
|
// actual url but used to finish a sentence.
|
|
|
|
|
if strings.HasSuffix(urlFull, ".") { |
|
|
|
|
end-- |
|
|
|
|
urlFull = urlFull[:len(urlFull)-1] |
|
|
|
|
if hash != "" { |
|
|
|
|
hash = hash[:len(hash)-1] |
|
|
|
|
} else if subpath != "" { |
|
|
|
|
subpath = subpath[:len(subpath)-1] |
|
|
|
|
} |
|
|
|
|
ret, ok := anyHashPatternExtract(node.Data) |
|
|
|
|
if !ok { |
|
|
|
|
node = node.NextSibling |
|
|
|
|
continue |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
if subpath != "" { |
|
|
|
|
text += subpath |
|
|
|
|
text := base.ShortSha(ret.CommitID) |
|
|
|
|
if ret.SubPath != "" { |
|
|
|
|
text += ret.SubPath |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
if hash != "" { |
|
|
|
|
text += " (" + hash + ")" |
|
|
|
|
if ret.QueryHash != "" { |
|
|
|
|
text += " (" + ret.QueryHash + ")" |
|
|
|
|
} |
|
|
|
|
replaceContent(node, start, end, createCodeLink(urlFull, text, "commit")) |
|
|
|
|
replaceContent(node, ret.PosStart, ret.PosEnd, createCodeLink(ret.FullURL, text, "commit")) |
|
|
|
|
node = node.NextSibling.NextSibling |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
@ -1021,19 +1034,16 @@ func comparePatternProcessor(ctx *RenderContext, node *html.Node) { |
|
|
|
|
if ctx.Metas == nil { |
|
|
|
|
return |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
next := node.NextSibling |
|
|
|
|
for node != nil && node != next { |
|
|
|
|
m := comparePattern.FindStringSubmatchIndex(node.Data) |
|
|
|
|
if m == nil { |
|
|
|
|
return |
|
|
|
|
nodeStop := node.NextSibling |
|
|
|
|
for node != nodeStop { |
|
|
|
|
if node.Type != html.TextNode { |
|
|
|
|
node = node.NextSibling |
|
|
|
|
continue |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
// Ensure that every group (m[0]...m[7]) has a match
|
|
|
|
|
for i := 0; i < 8; i++ { |
|
|
|
|
if m[i] == -1 { |
|
|
|
|
return |
|
|
|
|
} |
|
|
|
|
m := comparePattern.FindStringSubmatchIndex(node.Data) |
|
|
|
|
if m == nil || slices.Contains(m[:8], -1) { // ensure that every group (m[0]...m[7]) has a match
|
|
|
|
|
node = node.NextSibling |
|
|
|
|
continue |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
urlFull := node.Data[m[0]:m[1]] |
|
|
|
|