1
0
mirror of https://github.com/danog/blackfriday.git synced 2024-11-30 04:29:13 +01:00
blackfriday/markdown.go

844 lines
21 KiB
Go
Raw Normal View History

//
2011-06-28 04:11:32 +02:00
// Blackfriday Markdown Processor
// Available at http://github.com/russross/blackfriday
//
// Copyright © 2011 Russ Ross <russ@russross.com>.
// Distributed under the Simplified BSD License.
2011-06-28 04:11:32 +02:00
// See README.md for details.
//
2011-05-29 05:17:53 +02:00
//
//
// Markdown parsing and processing
//
//
2011-07-07 20:05:29 +02:00
// Blackfriday markdown processor.
//
// Translates plain text with simple formatting rules into HTML or LaTeX.
2011-05-29 05:17:53 +02:00
package blackfriday
2011-05-25 00:14:35 +02:00
import (
"bytes"
2012-03-08 05:36:31 +01:00
"unicode/utf8"
2011-05-25 00:14:35 +02:00
)
2011-09-26 23:45:49 +02:00
const VERSION = "1.1"
2011-05-29 05:17:53 +02:00
// These are the supported markdown parsing extensions.
// OR these values together to select multiple extensions.
2011-05-25 00:14:35 +02:00
const (
EXTENSION_NO_INTRA_EMPHASIS = 1 << iota // ignore emphasis markers inside words
EXTENSION_TABLES // render tables
EXTENSION_FENCED_CODE // render fenced code blocks
EXTENSION_AUTOLINK // detect embedded URLs that are not explicitly marked
EXTENSION_STRIKETHROUGH // strikethrough text using ~~test~~
EXTENSION_LAX_HTML_BLOCKS // loosen up HTML block parsing rules
EXTENSION_SPACE_HEADERS // be strict about prefix header rules
EXTENSION_HARD_LINE_BREAK // translate newlines into line breaks
EXTENSION_TAB_SIZE_EIGHT // expand tabs to eight spaces instead of four
EXTENSION_FOOTNOTES // Pandoc-style footnotes
EXTENSION_NO_EMPTY_LINE_BEFORE_BLOCK // No need to insert an empty line to start a (code, quote, order list, unorder list)block
EXTENSION_HEADER_IDS // specify header IDs with {#id}
2011-05-25 00:14:35 +02:00
)
2011-05-29 05:17:53 +02:00
// These are the possible flag values for the link renderer.
// Only a single one of these values will be used; they are not ORed together.
// These are mostly of interest if you are writing a new output format.
2011-05-25 00:14:35 +02:00
const (
2011-05-29 05:17:53 +02:00
LINK_TYPE_NOT_AUTOLINK = iota
LINK_TYPE_NORMAL
LINK_TYPE_EMAIL
2011-05-25 00:14:35 +02:00
)
2011-07-07 19:56:45 +02:00
// These are the possible flag values for the ListItem renderer.
2011-05-29 05:17:53 +02:00
// Multiple flag values may be ORed together.
// These are mostly of interest if you are writing a new output format.
2011-05-25 21:59:30 +02:00
const (
2011-05-29 05:17:53 +02:00
LIST_TYPE_ORDERED = 1 << iota
LIST_ITEM_CONTAINS_BLOCK
LIST_ITEM_BEGINNING_OF_LIST
2011-05-29 05:17:53 +02:00
LIST_ITEM_END_OF_LIST
2011-05-25 21:59:30 +02:00
)
2011-05-29 05:17:53 +02:00
// These are the possible flag values for the table cell renderer.
// Only a single one of these values will be used; they are not ORed together.
// These are mostly of interest if you are writing a new output format.
2011-05-25 21:59:30 +02:00
const (
2011-05-29 05:17:53 +02:00
TABLE_ALIGNMENT_LEFT = 1 << iota
TABLE_ALIGNMENT_RIGHT
TABLE_ALIGNMENT_CENTER = (TABLE_ALIGNMENT_LEFT | TABLE_ALIGNMENT_RIGHT)
2011-05-25 21:59:30 +02:00
)
2011-05-29 05:17:53 +02:00
// The size of a tab stop.
const (
TAB_SIZE_DEFAULT = 4
TAB_SIZE_EIGHT = 8
)
2011-05-29 05:17:53 +02:00
// These are the tags that are recognized as HTML block tags.
// Any of these can be included in markdown text without special escaping.
2011-06-29 00:02:12 +02:00
var blockTags = map[string]bool{
2011-05-28 17:49:21 +02:00
"p": true,
"dl": true,
"h1": true,
"h2": true,
"h3": true,
"h4": true,
"h5": true,
"h6": true,
"ol": true,
"ul": true,
"del": true,
"div": true,
"ins": true,
"pre": true,
"form": true,
"math": true,
"table": true,
"iframe": true,
"script": true,
"fieldset": true,
"noscript": true,
"blockquote": true,
2012-10-22 06:28:31 +02:00
// HTML5
"video": true,
"aside": true,
"canvas": true,
"figure": true,
"footer": true,
"header": true,
"hgroup": true,
"output": true,
"article": true,
"section": true,
"progress": true,
"figcaption": true,
2011-05-25 00:14:35 +02:00
}
2011-07-07 19:56:45 +02:00
// Renderer is the rendering interface.
2011-05-29 05:17:53 +02:00
// This is mostly of interest if you are implementing a new rendering format.
2011-07-07 19:56:45 +02:00
//
// When a byte slice is provided, it contains the (rendered) contents of the
// element.
//
// When a callback is provided instead, it will write the contents of the
// respective element directly to the output buffer and return true on success.
// If the callback returns false, the rendering function should reset the
// output buffer as though it had never been called.
//
2011-06-29 19:21:46 +02:00
// Currently Html and Latex implementations are provided
2011-06-29 19:13:17 +02:00
type Renderer interface {
// block-level callbacks
BlockCode(out *bytes.Buffer, text []byte, lang string)
BlockQuote(out *bytes.Buffer, text []byte)
BlockHtml(out *bytes.Buffer, text []byte)
Header(out *bytes.Buffer, text func() bool, level int, id string)
2011-06-29 19:13:17 +02:00
HRule(out *bytes.Buffer)
List(out *bytes.Buffer, text func() bool, flags int)
ListItem(out *bytes.Buffer, text []byte, flags int)
Paragraph(out *bytes.Buffer, text func() bool)
Table(out *bytes.Buffer, header []byte, body []byte, columnData []int)
TableRow(out *bytes.Buffer, text []byte)
TableHeaderCell(out *bytes.Buffer, text []byte, flags int)
2011-06-29 19:13:17 +02:00
TableCell(out *bytes.Buffer, text []byte, flags int)
Footnotes(out *bytes.Buffer, text func() bool)
FootnoteItem(out *bytes.Buffer, name, text []byte, flags int)
2011-06-29 19:13:17 +02:00
2011-06-29 21:00:54 +02:00
// Span-level callbacks
AutoLink(out *bytes.Buffer, link []byte, kind int)
CodeSpan(out *bytes.Buffer, text []byte)
DoubleEmphasis(out *bytes.Buffer, text []byte)
Emphasis(out *bytes.Buffer, text []byte)
Image(out *bytes.Buffer, link []byte, title []byte, alt []byte)
LineBreak(out *bytes.Buffer)
Link(out *bytes.Buffer, link []byte, title []byte, content []byte)
RawHtmlTag(out *bytes.Buffer, tag []byte)
TripleEmphasis(out *bytes.Buffer, text []byte)
StrikeThrough(out *bytes.Buffer, text []byte)
FootnoteRef(out *bytes.Buffer, ref []byte, id int)
2011-06-29 19:13:17 +02:00
// Low-level callbacks
Entity(out *bytes.Buffer, entity []byte)
NormalText(out *bytes.Buffer, text []byte)
// Header and footer
2011-06-29 19:13:17 +02:00
DocumentHeader(out *bytes.Buffer)
DocumentFooter(out *bytes.Buffer)
GetFlags() int
2011-05-25 00:14:35 +02:00
}
2011-06-29 19:21:46 +02:00
// Callback functions for inline parsing. One such function is defined
// for each character that triggers a response when parsing inline data.
2011-07-07 19:56:45 +02:00
type inlineParser func(p *parser, out *bytes.Buffer, data []byte, offset int) int
2011-05-26 04:46:16 +02:00
2011-07-07 19:56:45 +02:00
// Parser holds runtime state used by the parser.
// This is constructed by the Markdown function.
type parser struct {
2011-07-05 22:22:21 +02:00
r Renderer
refs map[string]*reference
inlineCallback [256]inlineParser
flags int
nesting int
maxNesting int
insideLink bool
// Footnotes need to be ordered as well as available to quickly check for
// presence. If a ref is also a footnote, it's stored both in refs and here
// in notes. Slice is nil if footnotes not enabled.
notes []*reference
2011-05-25 00:14:35 +02:00
}
2011-05-29 05:17:53 +02:00
//
//
// Public interface
//
2011-05-26 17:47:41 +02:00
//
2011-05-25 00:14:35 +02:00
2011-07-07 19:56:45 +02:00
// MarkdownBasic is a convenience function for simple rendering.
// It processes markdown input with no extensions enabled.
func MarkdownBasic(input []byte) []byte {
// set up the HTML renderer
htmlFlags := HTML_USE_XHTML
renderer := HtmlRenderer(htmlFlags, "", "")
// set up the parser
extensions := 0
return Markdown(input, renderer, extensions)
}
// Call Markdown with most useful extensions enabled
2011-07-07 19:56:45 +02:00
// MarkdownCommon is a convenience function for simple rendering.
// It processes markdown input with common extensions enabled, including:
//
// * Smartypants processing with smart fractions and LaTeX dashes
//
2013-04-14 00:44:18 +02:00
// * Intra-word emphasis suppression
2011-07-07 19:56:45 +02:00
//
// * Tables
//
// * Fenced code blocks
//
// * Autolinking
//
// * Strikethrough support
//
// * Strict header parsing
2014-04-05 21:45:57 +02:00
//
// * Custom Header IDs
func MarkdownCommon(input []byte) []byte {
// set up the HTML renderer
htmlFlags := 0
htmlFlags |= HTML_USE_XHTML
htmlFlags |= HTML_USE_SMARTYPANTS
htmlFlags |= HTML_SMARTYPANTS_FRACTIONS
htmlFlags |= HTML_SMARTYPANTS_LATEX_DASHES
htmlFlags |= HTML_SANITIZE_OUTPUT
renderer := HtmlRenderer(htmlFlags, "", "")
// set up the parser
extensions := 0
extensions |= EXTENSION_NO_INTRA_EMPHASIS
extensions |= EXTENSION_TABLES
extensions |= EXTENSION_FENCED_CODE
extensions |= EXTENSION_AUTOLINK
extensions |= EXTENSION_STRIKETHROUGH
extensions |= EXTENSION_SPACE_HEADERS
2014-04-05 21:45:57 +02:00
extensions |= EXTENSION_HEADER_IDS
return Markdown(input, renderer, extensions)
}
2011-07-07 19:56:45 +02:00
// Markdown is the main rendering function.
// It parses and renders a block of markdown-encoded text.
// The supplied Renderer is used to format the output, and extensions dictates
// which non-standard extensions are enabled.
//
// To use the supplied Html or LaTeX renderers, see HtmlRenderer and
// LatexRenderer, respectively.
2011-06-29 19:13:17 +02:00
func Markdown(input []byte, renderer Renderer, extensions int) []byte {
2011-05-29 05:17:53 +02:00
// no point in parsing if we can't render
if renderer == nil {
return nil
2011-05-26 04:46:16 +02:00
}
2011-05-29 05:17:53 +02:00
// fill in the render structure
2011-07-07 19:56:45 +02:00
p := new(parser)
p.r = renderer
p.flags = extensions
p.refs = make(map[string]*reference)
p.maxNesting = 16
p.insideLink = false
// register inline parsers
2011-07-07 19:56:45 +02:00
p.inlineCallback['*'] = emphasis
p.inlineCallback['_'] = emphasis
2011-06-29 19:13:17 +02:00
if extensions&EXTENSION_STRIKETHROUGH != 0 {
2011-07-07 19:56:45 +02:00
p.inlineCallback['~'] = emphasis
2011-05-29 05:17:53 +02:00
}
2011-07-07 19:56:45 +02:00
p.inlineCallback['`'] = codeSpan
p.inlineCallback['\n'] = lineBreak
p.inlineCallback['['] = link
p.inlineCallback['<'] = leftAngle
p.inlineCallback['\\'] = escape
p.inlineCallback['&'] = entity
2011-05-26 04:46:16 +02:00
2011-05-29 05:17:53 +02:00
if extensions&EXTENSION_AUTOLINK != 0 {
2011-07-07 19:56:45 +02:00
p.inlineCallback[':'] = autoLink
2011-05-29 05:17:53 +02:00
}
if extensions&EXTENSION_FOOTNOTES != 0 {
p.notes = make([]*reference, 0)
}
2011-07-07 19:56:45 +02:00
first := firstPass(p, input)
second := secondPass(p, first)
if renderer.GetFlags()&HTML_SANITIZE_OUTPUT != 0 {
second = sanitizeHtmlSafe(second)
}
return second
}
// first pass:
// - extract references
// - expand tabs
// - normalize newlines
// - copy everything else
// - add missing newlines before fenced code blocks
2011-07-07 19:56:45 +02:00
func firstPass(p *parser, input []byte) []byte {
var out bytes.Buffer
2011-06-29 00:02:12 +02:00
tabSize := TAB_SIZE_DEFAULT
2011-07-07 19:56:45 +02:00
if p.flags&EXTENSION_TAB_SIZE_EIGHT != 0 {
2011-06-29 00:02:12 +02:00
tabSize = TAB_SIZE_EIGHT
}
2011-05-29 05:17:53 +02:00
beg, end := 0, 0
lastLineWasBlank := false
lastFencedCodeBlockEnd := 0
2011-05-29 05:17:53 +02:00
for beg < len(input) { // iterate over lines
if end = isReference(p, input[beg:], tabSize); end > 0 {
2011-05-29 05:17:53 +02:00
beg += end
} else { // skip to the next line
end = beg
for end < len(input) && input[end] != '\n' && input[end] != '\r' {
end++
}
if p.flags&EXTENSION_FENCED_CODE != 0 {
// when last line was none blank and a fenced code block comes after
if beg >= lastFencedCodeBlockEnd {
// tmp var so we don't modify beyond bounds of `input`
var tmp = make([]byte, len(input[beg:]), len(input[beg:])+1)
copy(tmp, input[beg:])
if i := p.fencedCode(&out, append(tmp, '\n'), false); i > 0 {
if !lastLineWasBlank {
out.WriteByte('\n') // need to inject additional linebreak
}
lastFencedCodeBlockEnd = beg + i
}
}
lastLineWasBlank = end == beg
}
2011-05-29 05:17:53 +02:00
// add the line body if present
if end > beg {
if end < lastFencedCodeBlockEnd { // Do not expand tabs while inside fenced code blocks.
out.Write(input[beg:end])
} else {
expandTabs(&out, input[beg:end], tabSize)
}
2011-05-29 05:17:53 +02:00
}
out.WriteByte('\n')
2011-05-29 05:17:53 +02:00
if end < len(input) && input[end] == '\r' {
end++
}
if end < len(input) && input[end] == '\n' {
2011-05-29 05:17:53 +02:00
end++
}
beg = end
}
}
2011-07-03 18:51:07 +02:00
// empty input?
if out.Len() == 0 {
out.WriteByte('\n')
}
return out.Bytes()
}
2011-05-29 05:17:53 +02:00
// second pass: actual rendering
2011-07-07 19:56:45 +02:00
func secondPass(p *parser, input []byte) []byte {
var output bytes.Buffer
2011-05-29 05:17:53 +02:00
2011-07-07 19:56:45 +02:00
p.r.DocumentHeader(&output)
p.block(&output, input)
if p.flags&EXTENSION_FOOTNOTES != 0 && len(p.notes) > 0 {
p.r.Footnotes(&output, func() bool {
flags := LIST_ITEM_BEGINNING_OF_LIST
for _, ref := range p.notes {
var buf bytes.Buffer
if ref.hasBlock {
flags |= LIST_ITEM_CONTAINS_BLOCK
p.block(&buf, ref.title)
} else {
p.inline(&buf, ref.title)
}
p.r.FootnoteItem(&output, ref.link, buf.Bytes(), flags)
flags &^= LIST_ITEM_BEGINNING_OF_LIST | LIST_ITEM_CONTAINS_BLOCK
}
return true
})
}
2011-07-07 19:56:45 +02:00
p.r.DocumentFooter(&output)
2011-05-29 05:17:53 +02:00
2011-07-07 19:56:45 +02:00
if p.nesting != 0 {
2011-05-29 05:17:53 +02:00
panic("Nesting level did not end at zero")
}
return output.Bytes()
2011-05-29 05:17:53 +02:00
}
2011-05-25 21:59:30 +02:00
//
2011-05-29 05:17:53 +02:00
// Link references
2011-05-25 21:59:30 +02:00
//
2011-05-29 05:17:53 +02:00
// This section implements support for references that (usually) appear
// as footnotes in a document, and can be referenced anywhere in the document.
// The basic format is:
2011-05-25 21:59:30 +02:00
//
2011-05-29 05:17:53 +02:00
// [1]: http://www.google.com/ "Google"
// [2]: http://www.github.com/ "Github"
2011-05-25 21:59:30 +02:00
//
2011-05-29 05:17:53 +02:00
// Anywhere in the document, the reference can be linked by referring to its
// label, i.e., 1 and 2 in this example, as in:
//
// This library is hosted on [Github][2], a git hosting site.
//
// Actual footnotes as specified in Pandoc and supported by some other Markdown
// libraries such as php-markdown are also taken care of. They look like this:
//
// This sentence needs a bit of further explanation.[^note]
//
// [^note]: This is the explanation.
//
// Footnotes should be placed at the end of the document in an ordered list.
// Inline footnotes such as:
//
// Inline footnotes^[Not supported.] also exist.
//
// are not yet supported.
2011-05-25 21:59:30 +02:00
2011-05-29 05:17:53 +02:00
// References are parsed and stored in this struct.
type reference struct {
link []byte
title []byte
noteId int // 0 if not a footnote ref
hasBlock bool
2011-05-25 00:14:35 +02:00
}
2011-05-29 05:17:53 +02:00
// Check whether or not data starts with a reference link.
// If so, it is parsed and stored in the list of references
// (in the render struct).
// Returns the number of bytes to skip to move past it,
// or zero if the first line is not a reference.
func isReference(p *parser, data []byte, tabSize int) int {
2011-05-29 05:17:53 +02:00
// up to 3 optional leading spaces
if len(data) < 4 {
return 0
2011-05-25 00:14:35 +02:00
}
2011-05-29 05:17:53 +02:00
i := 0
for i < 3 && data[i] == ' ' {
i++
2011-05-25 00:14:35 +02:00
}
2011-05-29 05:17:53 +02:00
noteId := 0
2011-05-29 05:17:53 +02:00
// id part: anything but a newline between brackets
if data[i] != '[' {
return 0
2011-05-25 00:14:35 +02:00
}
2011-05-29 05:17:53 +02:00
i++
if p.flags&EXTENSION_FOOTNOTES != 0 {
if data[i] == '^' {
// we can set it to anything here because the proper noteIds will
// be assigned later during the second pass. It just has to be != 0
noteId = 1
i++
}
}
2011-06-29 00:02:12 +02:00
idOffset := i
2011-05-29 05:17:53 +02:00
for i < len(data) && data[i] != '\n' && data[i] != '\r' && data[i] != ']' {
i++
2011-05-25 00:14:35 +02:00
}
2011-05-29 05:17:53 +02:00
if i >= len(data) || data[i] != ']' {
return 0
2011-05-25 00:14:35 +02:00
}
2011-06-29 00:02:12 +02:00
idEnd := i
2011-05-25 00:14:35 +02:00
2011-05-29 05:17:53 +02:00
// spacer: colon (space | tab)* newline? (space | tab)*
i++
if i >= len(data) || data[i] != ':' {
return 0
2011-05-25 21:59:30 +02:00
}
2011-05-29 05:17:53 +02:00
i++
for i < len(data) && (data[i] == ' ' || data[i] == '\t') {
i++
2011-05-25 21:59:30 +02:00
}
2011-05-29 05:17:53 +02:00
if i < len(data) && (data[i] == '\n' || data[i] == '\r') {
i++
if i < len(data) && data[i] == '\n' && data[i-1] == '\r' {
i++
2011-05-25 21:59:30 +02:00
}
}
2011-05-29 05:17:53 +02:00
for i < len(data) && (data[i] == ' ' || data[i] == '\t') {
i++
}
if i >= len(data) {
return 0
2011-05-25 21:59:30 +02:00
}
var (
linkOffset, linkEnd int
titleOffset, titleEnd int
lineEnd int
raw []byte
hasBlock bool
)
if p.flags&EXTENSION_FOOTNOTES != 0 && noteId != 0 {
linkOffset, linkEnd, raw, hasBlock = scanFootnote(p, data, i, tabSize)
lineEnd = linkEnd
} else {
linkOffset, linkEnd, titleOffset, titleEnd, lineEnd = scanLinkRef(p, data, i)
}
if lineEnd == 0 {
return 0
}
// a valid ref has been found
ref := &reference{
noteId: noteId,
hasBlock: hasBlock,
}
if noteId > 0 {
// reusing the link field for the id since footnotes don't have links
ref.link = data[idOffset:idEnd]
// if footnote, it's not really a title, it's the contained text
ref.title = raw
} else {
ref.link = data[linkOffset:linkEnd]
ref.title = data[titleOffset:titleEnd]
}
// id matches are case-insensitive
id := string(bytes.ToLower(data[idOffset:idEnd]))
p.refs[id] = ref
return lineEnd
}
func scanLinkRef(p *parser, data []byte, i int) (linkOffset, linkEnd, titleOffset, titleEnd, lineEnd int) {
2011-05-29 05:17:53 +02:00
// link: whitespace-free sequence, optionally between angle brackets
if data[i] == '<' {
2011-05-26 00:00:01 +02:00
i++
}
linkOffset = i
2011-05-29 05:17:53 +02:00
for i < len(data) && data[i] != ' ' && data[i] != '\t' && data[i] != '\n' && data[i] != '\r' {
i++
2011-05-27 21:38:10 +02:00
}
linkEnd = i
2011-06-29 00:02:12 +02:00
if data[linkOffset] == '<' && data[linkEnd-1] == '>' {
linkOffset++
linkEnd--
2011-05-27 21:38:10 +02:00
}
2011-05-29 05:17:53 +02:00
// optional spacer: (space | tab)* (newline | '\'' | '"' | '(' )
for i < len(data) && (data[i] == ' ' || data[i] == '\t') {
i++
}
2011-05-29 05:17:53 +02:00
if i < len(data) && data[i] != '\n' && data[i] != '\r' && data[i] != '\'' && data[i] != '"' && data[i] != '(' {
return
}
2011-05-29 05:17:53 +02:00
// compute end-of-line
if i >= len(data) || data[i] == '\r' || data[i] == '\n' {
2011-06-29 00:02:12 +02:00
lineEnd = i
}
2011-05-29 05:17:53 +02:00
if i+1 < len(data) && data[i] == '\r' && data[i+1] == '\n' {
2011-06-29 00:02:12 +02:00
lineEnd++
2011-05-29 01:37:18 +02:00
}
2011-05-29 05:17:53 +02:00
// optional (space|tab)* spacer after a newline
2011-06-29 00:02:12 +02:00
if lineEnd > 0 {
i = lineEnd + 1
2011-05-29 05:17:53 +02:00
for i < len(data) && (data[i] == ' ' || data[i] == '\t') {
i++
2011-05-29 01:37:18 +02:00
}
}
2011-05-29 05:17:53 +02:00
// optional title: any non-newline sequence enclosed in '"() alone on its line
if i+1 < len(data) && (data[i] == '\'' || data[i] == '"' || data[i] == '(') {
i++
2011-06-29 00:02:12 +02:00
titleOffset = i
2011-05-29 05:17:53 +02:00
// look for EOL
for i < len(data) && data[i] != '\n' && data[i] != '\r' {
i++
}
2011-05-29 05:17:53 +02:00
if i+1 < len(data) && data[i] == '\n' && data[i+1] == '\r' {
2011-06-29 00:02:12 +02:00
titleEnd = i + 1
2011-05-29 05:17:53 +02:00
} else {
2011-06-29 00:02:12 +02:00
titleEnd = i
2011-05-29 01:37:18 +02:00
}
2011-05-29 05:17:53 +02:00
// step back
i--
2011-06-29 00:02:12 +02:00
for i > titleOffset && (data[i] == ' ' || data[i] == '\t') {
2011-05-29 05:17:53 +02:00
i--
2011-05-29 01:37:18 +02:00
}
2011-06-29 00:02:12 +02:00
if i > titleOffset && (data[i] == '\'' || data[i] == '"' || data[i] == ')') {
lineEnd = titleEnd
titleEnd = i
2011-05-29 01:37:18 +02:00
}
}
return
}
// The first bit of this logic is the same as (*parser).listItem, but the rest
// is much simpler. This function simply finds the entire block and shifts it
// over by one tab if it is indeed a block (just returns the line if it's not).
// blockEnd is the end of the section in the input buffer, and contents is the
// extracted text that was shifted over one tab. It will need to be rendered at
// the end of the document.
func scanFootnote(p *parser, data []byte, i, indentSize int) (blockStart, blockEnd int, contents []byte, hasBlock bool) {
2013-07-08 08:54:25 +02:00
if i == 0 || len(data) == 0 {
return
2011-05-29 01:37:18 +02:00
}
// skip leading whitespace on first line
2013-07-08 08:54:25 +02:00
for i < len(data) && data[i] == ' ' {
i++
}
blockStart = i
// find the end of the line
blockEnd = i
for i < len(data) && data[i-1] != '\n' {
i++
}
2011-05-29 01:37:18 +02:00
// get working buffer
var raw bytes.Buffer
// put the first line into the working buffer
raw.Write(data[blockEnd:i])
blockEnd = i
// process the following lines
containsBlankLine := false
gatherLines:
for blockEnd < len(data) {
i++
// find the end of this line
for i < len(data) && data[i-1] != '\n' {
i++
}
// if it is an empty line, guess that it is part of this item
// and move on to the next line
if p.isEmpty(data[blockEnd:i]) > 0 {
containsBlankLine = true
blockEnd = i
continue
}
n := 0
if n = isIndented(data[blockEnd:i], indentSize); n == 0 {
// this is the end of the block.
// we don't want to include this last line in the index.
break gatherLines
}
// if there were blank lines before this one, insert a new one now
if containsBlankLine {
raw.WriteByte('\n')
containsBlankLine = false
}
// get rid of that first tab, write to buffer
raw.Write(data[blockEnd+n : i])
hasBlock = true
blockEnd = i
}
if data[blockEnd-1] != '\n' {
raw.WriteByte('\n')
}
contents = raw.Bytes()
return
2011-05-29 01:37:18 +02:00
}
2011-05-29 05:17:53 +02:00
//
//
// Miscellaneous helper functions
//
//
2011-05-29 01:37:18 +02:00
2011-05-29 05:17:53 +02:00
// Test if a character is a punctuation symbol.
// Taken from a private function in regexp in the stdlib.
func ispunct(c byte) bool {
for _, r := range []byte("!\"#$%&'()*+,-./:;<=>?@[\\]^_`{|}~") {
if c == r {
return true
2011-05-29 01:37:18 +02:00
}
}
2011-05-29 05:17:53 +02:00
return false
2011-05-29 01:37:18 +02:00
}
2011-05-29 05:17:53 +02:00
// Test if a character is a whitespace character.
func isspace(c byte) bool {
return c == ' ' || c == '\t' || c == '\n' || c == '\r' || c == '\f' || c == '\v'
2011-05-29 01:37:18 +02:00
}
2013-08-09 11:24:26 +02:00
// Test if a character is letter.
func isletter(c byte) bool {
return (c >= 'a' && c <= 'z') || (c >= 'A' && c <= 'Z')
}
2011-05-29 05:17:53 +02:00
// Test if a character is a letter or a digit.
// TODO: check when this is looking for ASCII alnum and when it should use unicode
2011-05-29 05:17:53 +02:00
func isalnum(c byte) bool {
2013-08-09 11:24:26 +02:00
return (c >= '0' && c <= '9') || isletter(c)
2011-05-29 01:37:18 +02:00
}
2011-05-29 05:17:53 +02:00
// Replace tab characters with spaces, aligning to the next TAB_SIZE column.
// always ends output with a newline
2011-06-29 00:02:12 +02:00
func expandTabs(out *bytes.Buffer, line []byte, tabSize int) {
// first, check for common cases: no tabs, or only tabs at beginning of line
i, prefix := 0, 0
slowcase := false
for i = 0; i < len(line); i++ {
if line[i] == '\t' {
if prefix == i {
prefix++
} else {
slowcase = true
break
}
}
}
2011-05-26 04:46:16 +02:00
// no need to decode runes if all tabs are at the beginning of the line
if !slowcase {
2011-06-29 00:02:12 +02:00
for i = 0; i < prefix*tabSize; i++ {
out.WriteByte(' ')
}
out.Write(line[prefix:])
return
}
// the slow case: we need to count runes to figure out how
// many spaces to insert for each tab
column := 0
i = 0
2011-05-26 04:46:16 +02:00
for i < len(line) {
start := i
2011-05-26 04:46:16 +02:00
for i < len(line) && line[i] != '\t' {
_, size := utf8.DecodeRune(line[i:])
i += size
column++
2011-05-26 04:46:16 +02:00
}
if i > start {
out.Write(line[start:i])
2011-05-26 04:46:16 +02:00
}
if i >= len(line) {
break
}
for {
out.WriteByte(' ')
column++
2011-06-29 00:02:12 +02:00
if column%tabSize == 0 {
2011-05-26 04:46:16 +02:00
break
}
}
i++
}
}
// Find if a line counts as indented or not.
// Returns number of characters the indent is (0 = not indented).
func isIndented(data []byte, indentSize int) int {
if len(data) == 0 {
return 0
}
if data[0] == '\t' {
return 1
}
if len(data) < indentSize {
return 0
}
for i := 0; i < indentSize; i++ {
if data[i] != ' ' {
return 0
}
}
return indentSize
}
// Create a url-safe slug for fragments
func slugify(in []byte) []byte {
if len(in) == 0 {
return in
}
out := make([]byte, 0, len(in))
sym := false
for _, ch := range in {
if isalnum(ch) {
sym = false
out = append(out, ch)
} else if sym {
continue
} else {
out = append(out, '-')
sym = true
}
}
var a, b int
var ch byte
for a, ch = range out {
if ch != '-' {
break
}
}
for b = len(out) - 1; b > 0; b-- {
if out[b] != '-' {
break
}
}
return out[a : b+1]
}