Big refactor of how source files are used. Also added default destination extension...
authorspf13 <steve.francia@gmail.com>
Fri, 17 Oct 2014 00:20:09 +0000 (20:20 -0400)
committerspf13 <steve.francia@gmail.com>
Fri, 17 Oct 2014 00:20:09 +0000 (20:20 -0400)
24 files changed:
commands/convert.go
commands/hugo.go
create/content.go
docs/layouts/partials/menu.html
helpers/content.go [new file with mode: 0644]
helpers/general.go
helpers/path.go
hugolib/page.go
hugolib/pageGroup.go
hugolib/page_permalink_test.go
hugolib/page_test.go
hugolib/path_seperators_test.go
hugolib/permalinks.go
hugolib/planner.go
hugolib/shortcode.go
hugolib/site.go
hugolib/site_show_plan_test.go
hugolib/site_test.go
hugolib/site_url_test.go
hugolib/summary.go
source/file.go [new file with mode: 0644]
source/filesystem.go
source/filesystem_test.go
source/inmemory.go

index c9482bea426ee1ce0d0f76e6c549afcd0a951076..434c6896f84ae254ce5bf7d7e4a9aa29746496e4 100644 (file)
@@ -99,20 +99,20 @@ func convertContents(mark rune) (err error) {
 
        jww.FEEDBACK.Println("processing", len(site.Source.Files()), "content files")
        for _, file := range site.Source.Files() {
-               jww.INFO.Println("Attempting to convert", file.LogicalName)
-               page, err := hugolib.NewPage(file.LogicalName)
+               jww.INFO.Println("Attempting to convert", file.LogicalName())
+               page, err := hugolib.NewPage(file.LogicalName())
                if err != nil {
                        return err
                }
 
                psr, err := parser.ReadFrom(file.Contents)
                if err != nil {
-                       jww.ERROR.Println("Error processing file:", path.Join(file.Dir, file.LogicalName))
+                       jww.ERROR.Println("Error processing file:", file.Path())
                        return err
                }
                metadata, err := psr.Metadata()
                if err != nil {
-                       jww.ERROR.Println("Error processing file:", path.Join(file.Dir, file.LogicalName))
+                       jww.ERROR.Println("Error processing file:", file.Path())
                        return err
                }
 
@@ -128,7 +128,7 @@ func convertContents(mark rune) (err error) {
                        metadata = newmetadata
                }
 
-               page.Dir = file.Dir
+               //page.Dir = file.Dir
                page.SetSourceContent(psr.Content())
                page.SetSourceMetaData(metadata, mark)
 
index 1d3bdcd93afea40fe8590dd494d303ab3cb32011..7ebd6371b8cf1789a1263ec142abcc2cf5ff570b 100644 (file)
@@ -118,6 +118,7 @@ func InitializeConfig() {
        viper.SetDefault("Permalinks", make(hugolib.PermalinkOverrides, 0))
        viper.SetDefault("Sitemap", hugolib.Sitemap{Priority: -1})
        viper.SetDefault("PygmentsStyle", "monokai")
+       viper.SetDefault("DefaultExtension", "html")
        viper.SetDefault("PygmentsUseClasses", false)
        viper.SetDefault("DisableLiveReload", false)
        viper.SetDefault("PluralizeListTitles", true)
index 449d63b85d7ea17a6376fb7d93be944bcb1882d6..bb0f029dcb4fb77a42bebd9281d057dba7d0715b 100644 (file)
@@ -94,7 +94,7 @@ func NewContent(kind, name string) (err error) {
                newmetadata["date"] = time.Now().Format(time.RFC3339)
        }
 
-       page.Dir = viper.GetString("sourceDir")
+       //page.Dir = viper.GetString("sourceDir")
        page.SetSourceMetaData(newmetadata, parser.FormatToLeadRune(viper.GetString("MetaDataFormat")))
        page.SetSourceContent(psr.Content())
        if err = page.SafeSaveSourceAs(path.Join(viper.GetString("contentDir"), name)); err != nil {
index 5a9e1ddf11d6232c90ef985ce684c5860ecad302..4c0c53dbd4c499a714032656afd55bda355ad3fe 100644 (file)
@@ -31,7 +31,7 @@
           {{end}}
             <li> <a href="https://github.com/spf13/hugo/issues" target="blank"><i class='fa fa-life-ring'></i>Issues & Help</a> </li>
             {{ if .IsPage }}
-            {{ $File := .File }}  {{with $File.FileName }}<li><a href="https://github.com/spf13/hugo/edit/master/docs/content/{{ $File.Dir }}{{ $File.FileName }}" target="blank"><i class='fa fa-edit'></i> Refine this Page</a> </li>{{end}}
+            {{ $File := .File }}  {{with $File.Path }}<li><a href="https://github.com/spf13/hugo/edit/master/docs/content/{{ $File.Dir }}{{ $File.LogicalName }}" target="blank"><i class='fa fa-edit'></i> Refine this Page</a> </li>{{end}}
             {{ end }}
         </ul>
         <!-- sidebar menu end-->
diff --git a/helpers/content.go b/helpers/content.go
new file mode 100644 (file)
index 0000000..802187b
--- /dev/null
@@ -0,0 +1,231 @@
+// Copyright © 2014 Steve Francia <spf@spf13.com>.
+//
+// Licensed under the Simple Public License, Version 2.0 (the "License");
+// you may not use this file except in compliance with the License.
+// You may obtain a copy of the License at
+// http://opensource.org/licenses/Simple-2.0
+//
+// Unless required by applicable law or agreed to in writing, software
+// distributed under the License is distributed on an "AS IS" BASIS,
+// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+// See the License for the specific language governing permissions and
+// limitations under the License.
+
+package helpers
+
+import (
+       "bytes"
+       "html/template"
+       "os/exec"
+
+       "github.com/russross/blackfriday"
+       "github.com/spf13/viper"
+
+       jww "github.com/spf13/jwalterweatherman"
+
+       "strings"
+)
+
+var SummaryLength = 70
+var SummaryDivider = []byte("<!--more-->")
+
+func StripHTML(s string) string {
+       output := ""
+
+       // Shortcut strings with no tags in them
+       if !strings.ContainsAny(s, "<>") {
+               output = s
+       } else {
+               s = strings.Replace(s, "\n", " ", -1)
+               s = strings.Replace(s, "</p>", " \n", -1)
+               s = strings.Replace(s, "<br>", " \n", -1)
+               s = strings.Replace(s, "</br>", " \n", -1)
+
+               // Walk through the string removing all tags
+               b := new(bytes.Buffer)
+               inTag := false
+               for _, r := range s {
+                       switch r {
+                       case '<':
+                               inTag = true
+                       case '>':
+                               inTag = false
+                       default:
+                               if !inTag {
+                                       b.WriteRune(r)
+                               }
+                       }
+               }
+               output = b.String()
+       }
+       return output
+}
+
+func StripEmptyNav(in []byte) []byte {
+       return bytes.Replace(in, []byte("<nav>\n</nav>\n\n"), []byte(``), -1)
+}
+
+func BytesToHTML(b []byte) template.HTML {
+       return template.HTML(string(b))
+}
+
+func GetHtmlRenderer(defaultFlags int, footnoteref string) blackfriday.Renderer {
+       renderParameters := blackfriday.HtmlRendererParameters{
+               FootnoteAnchorPrefix:       viper.GetString("FootnoteAnchorPrefix"),
+               FootnoteReturnLinkContents: viper.GetString("FootnoteReturnLinkContents"),
+       }
+
+       if len(footnoteref) != 0 {
+               renderParameters.FootnoteAnchorPrefix = footnoteref + ":" +
+                       renderParameters.FootnoteAnchorPrefix
+       }
+
+       htmlFlags := defaultFlags
+       htmlFlags |= blackfriday.HTML_USE_XHTML
+       htmlFlags |= blackfriday.HTML_USE_SMARTYPANTS
+       htmlFlags |= blackfriday.HTML_SMARTYPANTS_FRACTIONS
+       htmlFlags |= blackfriday.HTML_SMARTYPANTS_LATEX_DASHES
+       htmlFlags |= blackfriday.HTML_FOOTNOTE_RETURN_LINKS
+
+       return blackfriday.HtmlRendererWithParameters(htmlFlags, "", "", renderParameters)
+}
+
+func GetMarkdownExtensions() int {
+       return 0 | blackfriday.EXTENSION_NO_INTRA_EMPHASIS |
+               blackfriday.EXTENSION_TABLES | blackfriday.EXTENSION_FENCED_CODE |
+               blackfriday.EXTENSION_AUTOLINK | blackfriday.EXTENSION_STRIKETHROUGH |
+               blackfriday.EXTENSION_SPACE_HEADERS | blackfriday.EXTENSION_FOOTNOTES |
+               blackfriday.EXTENSION_HEADER_IDS
+}
+
+func MarkdownRender(content []byte, footnoteref string) []byte {
+       return blackfriday.Markdown(content, GetHtmlRenderer(0, footnoteref),
+               GetMarkdownExtensions())
+}
+
+func MarkdownRenderWithTOC(content []byte, footnoteref string) []byte {
+       return blackfriday.Markdown(content,
+               GetHtmlRenderer(blackfriday.HTML_TOC, footnoteref),
+               GetMarkdownExtensions())
+}
+
+func ExtractTOC(content []byte) (newcontent []byte, toc []byte) {
+       origContent := make([]byte, len(content))
+       copy(origContent, content)
+       first := []byte(`<nav>
+<ul>`)
+
+       last := []byte(`</ul>
+</nav>`)
+
+       replacement := []byte(`<nav id="TableOfContents">
+<ul>`)
+
+       startOfTOC := bytes.Index(content, first)
+
+       peekEnd := len(content)
+       if peekEnd > 70+startOfTOC {
+               peekEnd = 70 + startOfTOC
+       }
+
+       if startOfTOC < 0 {
+               return StripEmptyNav(content), toc
+       }
+       // Need to peek ahead to see if this nav element is actually the right one.
+       correctNav := bytes.Index(content[startOfTOC:peekEnd], []byte(`#toc_0`))
+       if correctNav < 0 { // no match found
+               return content, toc
+       }
+       lengthOfTOC := bytes.Index(content[startOfTOC:], last) + len(last)
+       endOfTOC := startOfTOC + lengthOfTOC
+
+       newcontent = append(content[:startOfTOC], content[endOfTOC:]...)
+       toc = append(replacement, origContent[startOfTOC+len(first):endOfTOC]...)
+       return
+}
+
+func RenderBytesWithTOC(content []byte, pagefmt string, footnoteref string) []byte {
+       switch pagefmt {
+       default:
+               return MarkdownRenderWithTOC(content, footnoteref)
+       case "markdown":
+               return MarkdownRenderWithTOC(content, footnoteref)
+       case "rst":
+               return []byte(GetRstContent(content))
+       }
+}
+
+func RenderBytes(content []byte, pagefmt string, footnoteref string) []byte {
+       switch pagefmt {
+       default:
+               return MarkdownRender(content, footnoteref)
+       case "markdown":
+               return MarkdownRender(content, footnoteref)
+       case "rst":
+               return []byte(GetRstContent(content))
+       }
+}
+
+func TotalWords(s string) int {
+       return len(strings.Fields(s))
+}
+
+func WordCount(s string) map[string]int {
+       m := make(map[string]int)
+       for _, f := range strings.Fields(s) {
+               m[f] += 1
+       }
+
+       return m
+}
+
+func RemoveSummaryDivider(content []byte) []byte {
+       return bytes.Replace(content, SummaryDivider, []byte(""), -1)
+}
+
+func TruncateWords(s string, max int) string {
+       words := strings.Fields(s)
+       if max > len(words) {
+               return strings.Join(words, " ")
+       }
+
+       return strings.Join(words[:max], " ")
+}
+
+func TruncateWordsToWholeSentence(s string, max int) string {
+       words := strings.Fields(s)
+       if max > len(words) {
+               return strings.Join(words, " ")
+       }
+
+       for counter, word := range words[max:] {
+               if strings.HasSuffix(word, ".") ||
+                       strings.HasSuffix(word, "?") ||
+                       strings.HasSuffix(word, ".\"") ||
+                       strings.HasSuffix(word, "!") {
+                       return strings.Join(words[:max+counter+1], " ")
+               }
+       }
+
+       return strings.Join(words[:max], " ")
+}
+
+func GetRstContent(content []byte) string {
+       cleanContent := bytes.Replace(content, SummaryDivider, []byte(""), 1)
+
+       cmd := exec.Command("rst2html.py", "--leave-comments")
+       cmd.Stdin = bytes.NewReader(cleanContent)
+       var out bytes.Buffer
+       cmd.Stdout = &out
+       if err := cmd.Run(); err != nil {
+               jww.ERROR.Println(err)
+       }
+
+       rstLines := strings.Split(out.String(), "\n")
+       for i, line := range rstLines {
+               if strings.HasPrefix(line, "<body>") {
+                       rstLines = (rstLines[i+1 : len(rstLines)-3])
+               }
+       }
+       return strings.Join(rstLines, "\n")
+}
index a2ebac9e1ca6931055c84e8907480534cd8aedd4..89e3857b0dff2734f0c1a786f2f9f632f47acd37 100644 (file)
@@ -15,43 +15,14 @@ package helpers
 
 import (
        "bytes"
+       "crypto/md5"
+       "encoding/hex"
        "fmt"
+       "io"
        "net"
        "strings"
 )
 
-func StripHTML(s string) string {
-       output := ""
-
-       // Shortcut strings with no tags in them
-       if !strings.ContainsAny(s, "<>") {
-               output = s
-       } else {
-               s = strings.Replace(s, "\n", " ", -1)
-               s = strings.Replace(s, "</p>", " \n", -1)
-               s = strings.Replace(s, "<br>", " \n", -1)
-               s = strings.Replace(s, "</br>", " \n", -1)
-
-               // Walk through the string removing all tags
-               b := new(bytes.Buffer)
-               inTag := false
-               for _, r := range s {
-                       switch r {
-                       case '<':
-                               inTag = true
-                       case '>':
-                               inTag = false
-                       default:
-                               if !inTag {
-                                       b.WriteRune(r)
-                               }
-                       }
-               }
-               output = b.String()
-       }
-       return output
-}
-
 func FindAvailablePort() (*net.TCPAddr, error) {
        l, err := net.Listen("tcp", ":0")
        if err == nil {
@@ -64,3 +35,42 @@ func FindAvailablePort() (*net.TCPAddr, error) {
        }
        return nil, err
 }
+
+func GuessType(in string) string {
+       switch strings.ToLower(in) {
+       case "md", "markdown", "mdown":
+               return "markdown"
+       case "rst":
+               return "rst"
+       case "html", "htm":
+               return "html"
+       }
+
+       return "unknown"
+}
+
+func ReaderToBytes(lines io.Reader) []byte {
+       b := new(bytes.Buffer)
+       b.ReadFrom(lines)
+       return b.Bytes()
+}
+
+// sliceToLower goes through the source slice and lowers all values.
+func SliceToLower(s []string) []string {
+       if s == nil {
+               return nil
+       }
+
+       l := make([]string, len(s))
+       for i, v := range s {
+               l[i] = strings.ToLower(v)
+       }
+
+       return l
+}
+
+func Md5String(f string) string {
+       h := md5.New()
+       h.Write([]byte(f))
+       return hex.EncodeToString(h.Sum([]byte{}))
+}
index 952c5428fc157c2c76cf4efc9ca383582457cb40..ace9432ace1b08b01fe921040ac6628e760635d4 100644 (file)
@@ -166,16 +166,52 @@ func FileAndExt(in string) (name string, ext string) {
        return
 }
 
+func GetRelativePath(path, base string) (final string, err error) {
+       if filepath.IsAbs(path) && base == "" {
+               return "", errors.New("source: missing base directory")
+       }
+       name := filepath.Clean(path)
+       base = filepath.Clean(base)
+
+       name, err = filepath.Rel(base, name)
+       if err != nil {
+               return "", err
+       }
+       name = filepath.ToSlash(name)
+       return name, nil
+}
+
+// Given a source path, determine the section
 func GuessSection(in string) string {
-       x := strings.Split(in, "/")
-       x = x[:len(x)-1]
-       if len(x) == 0 {
+       parts := strings.Split(in, "/")
+
+       if len(parts) == 0 {
                return ""
        }
-       if x[0] == "content" {
-               x = x[1:]
+
+       // trim filename
+       if !strings.HasSuffix(in, "/") {
+               parts = parts[:len(parts)-1]
        }
-       return path.Join(x...)
+
+       if len(parts) == 0 {
+               return ""
+       }
+
+       // if first directory is "content", return second directory
+       section := ""
+
+       if parts[0] == "content" && len(parts) > 1 {
+               section = parts[1]
+       } else {
+               section = parts[0]
+       }
+
+       if section == "." {
+               return ""
+       }
+
+       return section
 }
 
 func PathPrep(ugly bool, in string) string {
index 7cd40050482bae11409d773d8b7c7d0fd92ae532..f5b5a2cdbc6045b16ef9e568686f560c34d69802 100644 (file)
@@ -15,56 +15,63 @@ package hugolib
 
 import (
        "bytes"
-       "crypto/md5"
-       "encoding/hex"
        "errors"
        "fmt"
        "html/template"
        "io"
        "net/url"
        "path"
-       "path/filepath"
        "strings"
        "time"
 
-       "github.com/russross/blackfriday"
        "github.com/spf13/cast"
        "github.com/spf13/hugo/helpers"
        "github.com/spf13/hugo/parser"
+       "github.com/spf13/hugo/source"
        jww "github.com/spf13/jwalterweatherman"
        "github.com/spf13/viper"
 )
 
 type Page struct {
-       Status            string
-       Images            []string
-       rawContent        []byte
-       Content           template.HTML
-       Summary           template.HTML
-       TableOfContents   template.HTML
-       Truncated         bool
-       plain             string // TODO should be []byte
-       Params            map[string]interface{}
-       contentType       string
-       Draft             bool
-       PublishDate       time.Time
-       Aliases           []string
-       Tmpl              Template
-       Markup            string
-       renderable        bool
-       layout            string
-       linkTitle         string
-       frontmatter       []byte
-       sourceFrontmatter []byte
-       sourceContent     []byte
+       Params          map[string]interface{}
+       Content         template.HTML
+       Summary         template.HTML
+       Aliases         []string
+       Status          string
+       Images          []string
+       TableOfContents template.HTML
+       Truncated       bool
+       Draft           bool
+       PublishDate     time.Time
+       Tmpl            Template
+       Markup          string
+
+       extension   string
+       contentType string
+       renderable  bool
+       layout      string
+       linkTitle   string
+       frontmatter []byte
+       rawContent  []byte
+       plain       string // TODO should be []byte
+       //sourceFrontmatter []byte
+       //sourceContent     []byte
        PageMeta
-       File
+       //SourceFile source.File
+       Source
        Position
        Node
+       //Destination source.File
 }
 
-type File struct {
-       Name, FileName, Extension, Dir, UniqueId string
+//type File struct {
+//Name, FileName, Extension, Dir, UniqueId string
+//}
+
+type Source struct {
+       Frontmatter []byte
+       Content     []byte
+       source.File
 }
 
 type PageMeta struct {
@@ -97,75 +104,40 @@ func (p *Page) IsPage() bool {
 }
 
 func (p *Page) UniqueId() string {
-       return p.File.UniqueId
+       return p.Source.UniqueId()
 }
 
 func (p *Page) setSummary() {
-       if bytes.Contains(p.rawContent, summaryDivider) {
+       if bytes.Contains(p.rawContent, helpers.SummaryDivider) {
                // If user defines split:
                // Split then render
                p.Truncated = true // by definition
-               header := bytes.Split(p.rawContent, summaryDivider)[0]
-               p.Summary = bytesToHTML(p.renderBytes(header))
+               header := bytes.Split(p.rawContent, helpers.SummaryDivider)[0]
+               p.Summary = helpers.BytesToHTML(p.renderBytes(header))
        } else {
                // If hugo defines split:
                // render, strip html, then split
                plain := strings.TrimSpace(p.Plain())
-               p.Summary = bytesToHTML([]byte(TruncateWordsToWholeSentence(plain, summaryLength)))
+               p.Summary = helpers.BytesToHTML([]byte(helpers.TruncateWordsToWholeSentence(plain, helpers.SummaryLength)))
                p.Truncated = len(p.Summary) != len(plain)
        }
 }
 
-func stripEmptyNav(in []byte) []byte {
-       return bytes.Replace(in, []byte("<nav>\n</nav>\n\n"), []byte(``), -1)
-}
-
-func bytesToHTML(b []byte) template.HTML {
-       return template.HTML(string(b))
-}
-
 func (p *Page) renderBytes(content []byte) []byte {
-       return renderBytes(content, p.guessMarkupType(), p.UniqueId())
+       return helpers.RenderBytes(content, p.guessMarkupType(), p.UniqueId())
 }
 
 func (p *Page) renderContent(content []byte) []byte {
-       return renderBytesWithTOC(content, p.guessMarkupType(), p.UniqueId())
-}
-
-func renderBytesWithTOC(content []byte, pagefmt string, footnoteref string) []byte {
-       switch pagefmt {
-       default:
-               return markdownRenderWithTOC(content, footnoteref)
-       case "markdown":
-               return markdownRenderWithTOC(content, footnoteref)
-       case "rst":
-               return []byte(getRstContent(content))
-       }
-}
-
-func renderBytes(content []byte, pagefmt string, footnoteref string) []byte {
-       switch pagefmt {
-       default:
-               return markdownRender(content, footnoteref)
-       case "markdown":
-               return markdownRender(content, footnoteref)
-       case "rst":
-               return []byte(getRstContent(content))
-       }
+       return helpers.RenderBytesWithTOC(content, p.guessMarkupType(), p.UniqueId())
 }
 
 func newPage(filename string) *Page {
-       name := filepath.Base(filename)
-       // strip off the extension
-       name = name[:len(name)-len(filepath.Ext(name))]
-
        page := Page{contentType: "",
-               File:   File{Name: name, FileName: filename, Extension: "html", UniqueId: md5ForFilename(filename)},
+               Source: Source{File: *source.NewFile(filename)},
                Node:   Node{Keywords: []string{}, Sitemap: Sitemap{Priority: -1}},
                Params: make(map[string]interface{})}
 
-       jww.DEBUG.Println("Reading from", page.File.FileName)
-       page.guessSection()
+       jww.DEBUG.Println("Reading from", page.File.Path())
        return &page
 }
 
@@ -173,24 +145,22 @@ func (p *Page) IsRenderable() bool {
        return p.renderable
 }
 
-func (p *Page) guessSection() {
-       if p.Section == "" {
-               p.Section = helpers.GuessSection(p.FileName)
-       }
-}
-
 func (page *Page) Type() string {
        if page.contentType != "" {
                return page.contentType
        }
-       page.guessSection()
-       if x := page.Section; x != "" {
+
+       if x := page.Section(); x != "" {
                return x
        }
 
        return "page"
 }
 
+func (page *Page) Section() string {
+       return page.Source.Section()
+}
+
 func (page *Page) Layout(l ...string) []string {
        if page.layout != "" {
                return layouts(page.Type(), page.layout)
@@ -261,14 +231,14 @@ func (p *Page) ReadFrom(buf io.Reader) (err error) {
 }
 
 func (p *Page) analyzePage() {
-       p.WordCount = TotalWords(p.Plain())
+       p.WordCount = helpers.TotalWords(p.Plain())
        p.FuzzyWordCount = int((p.WordCount+100)/100) * 100
        p.ReadingTime = int((p.WordCount + 212) / 213)
 }
 
 func (p *Page) permalink() (*url.URL, error) {
        baseUrl := string(p.Site.BaseUrl)
-       dir := strings.TrimSpace(p.Dir)
+       dir := strings.TrimSpace(p.Source.Dir())
        pSlug := strings.TrimSpace(p.Slug)
        pUrl := strings.TrimSpace(p.Url)
        var permalink string
@@ -278,7 +248,7 @@ func (p *Page) permalink() (*url.URL, error) {
                return helpers.MakePermalink(baseUrl, pUrl), nil
        }
 
-       if override, ok := p.Site.Permalinks[p.Section]; ok {
+       if override, ok := p.Site.Permalinks[p.Section()]; ok {
                permalink, err = override.Expand(p)
 
                if err != nil {
@@ -287,16 +257,24 @@ func (p *Page) permalink() (*url.URL, error) {
                // fmt.Printf("have a section override for %q in section %s → %s\n", p.Title, p.Section, permalink)
        } else {
                if len(pSlug) > 0 {
-                       permalink = helpers.UrlPrep(viper.GetBool("UglyUrls"), path.Join(dir, p.Slug+"."+p.Extension))
+                       permalink = helpers.UrlPrep(viper.GetBool("UglyUrls"), path.Join(dir, p.Slug+"."+p.Extension()))
                } else {
-                       _, t := path.Split(p.FileName)
-                       permalink = helpers.UrlPrep(viper.GetBool("UglyUrls"), path.Join(dir, helpers.ReplaceExtension(strings.TrimSpace(t), p.Extension)))
+                       _, t := path.Split(p.Source.LogicalName())
+                       permalink = helpers.UrlPrep(viper.GetBool("UglyUrls"), path.Join(dir, helpers.ReplaceExtension(strings.TrimSpace(t), p.Extension())))
                }
        }
 
        return helpers.MakePermalink(baseUrl, permalink), nil
 }
 
+func (p *Page) Extension() string {
+       if p.extension != "" {
+               return p.extension
+       } else {
+               return viper.GetString("DefaultExtension")
+       }
+}
+
 func (p *Page) LinkTitle() string {
        if len(p.linkTitle) > 0 {
                return p.linkTitle
@@ -370,6 +348,8 @@ func (page *Page) update(f interface{}) error {
                        page.Url = helpers.Urlize(cast.ToString(v))
                case "type":
                        page.contentType = cast.ToString(v)
+               case "extension", "ext":
+                       page.extension = cast.ToString(v)
                case "keywords":
                        page.Keywords = cast.ToStringSlice(v)
                case "date":
@@ -445,7 +425,7 @@ func (page *Page) GetParam(key string) interface{} {
        case time.Time:
                return cast.ToTime(v)
        case []string:
-               return sliceToLower(v.([]string))
+               return helpers.SliceToLower(v.([]string))
        }
        return nil
 }
@@ -543,31 +523,13 @@ func (p *Page) Render(layout ...string) template.HTML {
 func (page *Page) guessMarkupType() string {
        // First try the explicitly set markup from the frontmatter
        if page.Markup != "" {
-               format := guessType(page.Markup)
+               format := helpers.GuessType(page.Markup)
                if format != "unknown" {
                        return format
                }
        }
 
-       // Then try to guess from the extension
-       ext := strings.ToLower(path.Ext(page.FileName))
-       if strings.HasPrefix(ext, ".") {
-               return guessType(ext[1:])
-       }
-
-       return "unknown"
-}
-
-func guessType(in string) string {
-       switch strings.ToLower(in) {
-       case "md", "markdown", "mdown":
-               return "markdown"
-       case "rst":
-               return "rst"
-       case "html", "htm":
-               return "html"
-       }
-       return "unknown"
+       return helpers.GuessType(page.Source.Ext())
 }
 
 func (page *Page) detectFrontMatter() (f *parser.FrontmatterType) {
@@ -585,7 +547,7 @@ func (page *Page) parse(reader io.Reader) error {
        meta, err := psr.Metadata()
        if meta != nil {
                if err != nil {
-                       jww.ERROR.Printf("Error parsing page meta data for %s", page.FileName)
+                       jww.ERROR.Printf("Error parsing page meta data for %s", page.File.Path())
                        jww.ERROR.Println(err)
                        return err
                }
@@ -601,7 +563,7 @@ func (page *Page) parse(reader io.Reader) error {
 }
 
 func (page *Page) SetSourceContent(content []byte) {
-       page.sourceContent = content
+       page.Source.Content = content
 }
 
 func (page *Page) SetSourceMetaData(in interface{}, mark rune) (err error) {
@@ -611,7 +573,7 @@ func (page *Page) SetSourceMetaData(in interface{}, mark rune) (err error) {
        }
        by = append(by, '\n')
 
-       page.sourceFrontmatter = by
+       page.Source.Frontmatter = by
 
        return nil
 }
@@ -626,8 +588,8 @@ func (page *Page) SaveSourceAs(path string) error {
 
 func (page *Page) saveSourceAs(path string, safe bool) error {
        b := new(bytes.Buffer)
-       b.Write(page.sourceFrontmatter)
-       b.Write(page.sourceContent)
+       b.Write(page.Source.Frontmatter)
+       b.Write(page.Source.Content)
 
        err := page.saveSource(b.Bytes(), path, safe)
        if err != nil {
@@ -666,100 +628,19 @@ func (page *Page) Convert() error {
        markupType := page.guessMarkupType()
        switch markupType {
        case "markdown", "rst":
-               tmpContent, tmpTableOfContents := extractTOC(page.renderContent(RemoveSummaryDivider(page.rawContent)))
-               page.Content = bytesToHTML(tmpContent)
-               page.TableOfContents = bytesToHTML(tmpTableOfContents)
+               tmpContent, tmpTableOfContents := helpers.ExtractTOC(page.renderContent(helpers.RemoveSummaryDivider(page.rawContent)))
+               page.Content = helpers.BytesToHTML(tmpContent)
+               page.TableOfContents = helpers.BytesToHTML(tmpTableOfContents)
        case "html":
-               page.Content = bytesToHTML(page.rawContent)
+               page.Content = helpers.BytesToHTML(page.rawContent)
        default:
-               return fmt.Errorf("Error converting unsupported file type '%s' for page '%s'", markupType, page.FileName)
+               return fmt.Errorf("Error converting unsupported file type '%s' for page '%s'", markupType, page.Source.Path())
        }
        return nil
 }
 
-func getHtmlRenderer(defaultFlags int, footnoteref string) blackfriday.Renderer {
-       renderParameters := blackfriday.HtmlRendererParameters{
-               FootnoteAnchorPrefix:       viper.GetString("FootnoteAnchorPrefix"),
-               FootnoteReturnLinkContents: viper.GetString("FootnoteReturnLinkContents"),
-       }
-
-       if len(footnoteref) != 0 {
-               renderParameters.FootnoteAnchorPrefix = footnoteref + ":" +
-                       renderParameters.FootnoteAnchorPrefix
-       }
-
-       htmlFlags := defaultFlags
-       htmlFlags |= blackfriday.HTML_USE_XHTML
-       htmlFlags |= blackfriday.HTML_USE_SMARTYPANTS
-       htmlFlags |= blackfriday.HTML_SMARTYPANTS_FRACTIONS
-       htmlFlags |= blackfriday.HTML_SMARTYPANTS_LATEX_DASHES
-       htmlFlags |= blackfriday.HTML_FOOTNOTE_RETURN_LINKS
-
-       return blackfriday.HtmlRendererWithParameters(htmlFlags, "", "", renderParameters)
-}
-
-func getMarkdownExtensions() int {
-       return 0 | blackfriday.EXTENSION_NO_INTRA_EMPHASIS |
-               blackfriday.EXTENSION_TABLES | blackfriday.EXTENSION_FENCED_CODE |
-               blackfriday.EXTENSION_AUTOLINK | blackfriday.EXTENSION_STRIKETHROUGH |
-               blackfriday.EXTENSION_SPACE_HEADERS | blackfriday.EXTENSION_FOOTNOTES |
-               blackfriday.EXTENSION_HEADER_IDS
-}
-
-func markdownRender(content []byte, footnoteref string) []byte {
-       return blackfriday.Markdown(content, getHtmlRenderer(0, footnoteref),
-               getMarkdownExtensions())
-}
-
-func markdownRenderWithTOC(content []byte, footnoteref string) []byte {
-       return blackfriday.Markdown(content,
-               getHtmlRenderer(blackfriday.HTML_TOC, footnoteref),
-               getMarkdownExtensions())
-}
-
-func extractTOC(content []byte) (newcontent []byte, toc []byte) {
-       origContent := make([]byte, len(content))
-       copy(origContent, content)
-       first := []byte(`<nav>
-<ul>`)
-
-       last := []byte(`</ul>
-</nav>`)
-
-       replacement := []byte(`<nav id="TableOfContents">
-<ul>`)
-
-       startOfTOC := bytes.Index(content, first)
-
-       peekEnd := len(content)
-       if peekEnd > 70+startOfTOC {
-               peekEnd = 70 + startOfTOC
-       }
-
-       if startOfTOC < 0 {
-               return stripEmptyNav(content), toc
-       }
-       // Need to peek ahead to see if this nav element is actually the right one.
-       correctNav := bytes.Index(content[startOfTOC:peekEnd], []byte(`#toc_0`))
-       if correctNav < 0 { // no match found
-               return content, toc
-       }
-       lengthOfTOC := bytes.Index(content[startOfTOC:], last) + len(last)
-       endOfTOC := startOfTOC + lengthOfTOC
-
-       newcontent = append(content[:startOfTOC], content[endOfTOC:]...)
-       toc = append(replacement, origContent[startOfTOC+len(first):endOfTOC]...)
-       return
-}
-
-func ReaderToBytes(lines io.Reader) []byte {
-       b := new(bytes.Buffer)
-       b.ReadFrom(lines)
-       return b.Bytes()
-}
-
 func (p *Page) FullFilePath() string {
-       return path.Join(p.Dir, p.FileName)
+       return path.Join(p.Source.Dir(), p.Source.Path())
 }
 
 func (p *Page) TargetPath() (outfile string) {
@@ -775,7 +656,7 @@ func (p *Page) TargetPath() (outfile string) {
        }
 
        // If there's a Permalink specification, we use that
-       if override, ok := p.Site.Permalinks[p.Section]; ok {
+       if override, ok := p.Site.Permalinks[p.Section()]; ok {
                var err error
                outfile, err = override.Expand(p)
                if err == nil {
@@ -787,32 +668,11 @@ func (p *Page) TargetPath() (outfile string) {
        }
 
        if len(strings.TrimSpace(p.Slug)) > 0 {
-               outfile = strings.TrimSpace(p.Slug) + "." + p.Extension
+               outfile = strings.TrimSpace(p.Slug) + "." + p.Extension()
        } else {
                // Fall back to filename
-               _, t := path.Split(p.FileName)
-               outfile = helpers.ReplaceExtension(strings.TrimSpace(t), p.Extension)
-       }
-
-       return path.Join(p.Dir, strings.TrimSpace(outfile))
-}
-
-// sliceToLower goes through the source slice and lowers all values.
-func sliceToLower(s []string) []string {
-       if s == nil {
-               return nil
+               outfile = helpers.ReplaceExtension(p.Source.LogicalName(), p.Extension())
        }
 
-       l := make([]string, len(s))
-       for i, v := range s {
-               l[i] = strings.ToLower(v)
-       }
-
-       return l
-}
-
-func md5ForFilename(f string) string {
-       h := md5.New()
-       h.Write([]byte(f))
-       return hex.EncodeToString(h.Sum([]byte{}))
+       return path.Join(p.Source.Dir(), strings.TrimSpace(outfile))
 }
index 9d8d9d62cf9ac2cf3be3a06c32352ec4e6a3a972..f859c2e508f3adad9b0fb6eaffe0d1f334787fc1 100644 (file)
@@ -83,20 +83,25 @@ func (p Pages) GroupBy(key string, order ...string) (PagesGroup, error) {
                direction = "desc"
        }
 
-       ppt := reflect.TypeOf(&Page{})
+       ppt := reflect.TypeOf(&Page{}) // *hugolib.Page
+
        ft, ok := ppt.Elem().FieldByName(key)
+
        if !ok {
                return nil, errors.New("No such field in Page struct")
        }
+
        tmp := reflect.MakeMap(reflect.MapOf(ft.Type, reflect.SliceOf(ppt)))
 
        for _, e := range p {
                ppv := reflect.ValueOf(e)
                fv := ppv.Elem().FieldByName(key)
-               if !tmp.MapIndex(fv).IsValid() {
-                       tmp.SetMapIndex(fv, reflect.MakeSlice(reflect.SliceOf(ppt), 0, 0))
+               if !fv.IsNil() {
+                       if !tmp.MapIndex(fv).IsValid() {
+                               tmp.SetMapIndex(fv, reflect.MakeSlice(reflect.SliceOf(ppt), 0, 0))
+                       }
+                       tmp.SetMapIndex(fv, reflect.Append(tmp.MapIndex(fv), ppv))
                }
-               tmp.SetMapIndex(fv, reflect.Append(tmp.MapIndex(fv), ppv))
        }
 
        var r []PageGroup
index b20362d5eb81d4229a37af8e96a1e69b82e2c9c4..f420b19c629619fdf192f8db566e811543b745db 100644 (file)
@@ -4,6 +4,7 @@ import (
        "html/template"
        "testing"
 
+       "github.com/spf13/hugo/source"
        "github.com/spf13/viper"
 )
 
@@ -33,6 +34,8 @@ func TestPermalink(t *testing.T) {
                {"x/y/z/boofar.md", "x/y/z", "", "", "/z/y/q/", false, "/z/y/q/", "/z/y/q/"},
        }
 
+       viper.Set("DefaultExtension", "html")
+
        for i, test := range tests {
                viper.Set("uglyurls", test.uglyurls)
                p := &Page{
@@ -45,7 +48,7 @@ func TestPermalink(t *testing.T) {
                                        BaseUrl: test.base,
                                },
                        },
-                       File: File{FileName: test.file, Dir: test.dir, Extension: "html"},
+                       Source: Source{File: *source.NewFile(test.file)},
                }
 
                if test.slug != "" {
index 67c1270f322d770253c01882f3146eaee86aabcc..6bb554bdf5b67505b00f837e498bdfc2fd961522 100644 (file)
@@ -6,6 +6,8 @@ import (
        "strings"
        "testing"
        "time"
+
+       "github.com/spf13/hugo/helpers"
 )
 
 var EMPTY_PAGE = ""
@@ -507,7 +509,7 @@ func TestDegenerateInvalidFrontMatterLeadingWhitespace(t *testing.T) {
 func TestSectionEvaluation(t *testing.T) {
        page, _ := NewPage("blue/file1.md")
        page.ReadFrom(strings.NewReader(SIMPLE_PAGE))
-       if page.Section != "blue" {
+       if page.Section() != "blue" {
                t.Errorf("Section should be %s, got: %s", "blue", page.Section)
        }
 }
@@ -529,12 +531,12 @@ func TestLayoutOverride(t *testing.T) {
                path           string
                expectedLayout []string
        }{
-               {SIMPLE_PAGE_NOLAYOUT, path_content_two_dir, L("dub/sub/single.html", "dub/single.html", "_default/single.html")},
+               {SIMPLE_PAGE_NOLAYOUT, path_content_two_dir, L("dub/single.html", "_default/single.html")},
                {SIMPLE_PAGE_NOLAYOUT, path_content_one_dir, L("gub/single.html", "_default/single.html")},
                {SIMPLE_PAGE_NOLAYOUT, path_content_no_dir, L("page/single.html", "_default/single.html")},
                {SIMPLE_PAGE_NOLAYOUT, path_one_directory, L("fub/single.html", "_default/single.html")},
                {SIMPLE_PAGE_NOLAYOUT, path_no_directory, L("page/single.html", "_default/single.html")},
-               {SIMPLE_PAGE_LAYOUT_FOOBAR, path_content_two_dir, L("dub/sub/foobar.html", "dub/foobar.html", "_default/foobar.html")},
+               {SIMPLE_PAGE_LAYOUT_FOOBAR, path_content_two_dir, L("dub/foobar.html", "_default/foobar.html")},
                {SIMPLE_PAGE_LAYOUT_FOOBAR, path_content_one_dir, L("gub/foobar.html", "_default/foobar.html")},
                {SIMPLE_PAGE_LAYOUT_FOOBAR, path_one_directory, L("fub/foobar.html", "_default/foobar.html")},
                {SIMPLE_PAGE_LAYOUT_FOOBAR, path_no_directory, L("page/foobar.html", "_default/foobar.html")},
@@ -576,7 +578,7 @@ func TestSliceToLower(t *testing.T) {
        }
 
        for _, test := range tests {
-               res := sliceToLower(test.value)
+               res := helpers.SliceToLower(test.value)
                for i, val := range res {
                        if val != test.expected[i] {
                                t.Errorf("Case mismatch. Expected %s, got %s", test.expected[i], res[i])
index 187c155b5e4e2d7eb20aa3e20e950711d358d131..589619d8cd76c110e6531add23ecb7f8c4cdcaff 100644 (file)
@@ -17,7 +17,7 @@ func TestDegenerateMissingFolderInPageFilename(t *testing.T) {
        if err != nil {
                t.Fatalf("Error in NewPageFrom")
        }
-       if p.Section != "" {
+       if p.Section() != "" {
                t.Fatalf("No section should be set for a file path: foobar")
        }
 }
@@ -31,17 +31,16 @@ func TestNewPageWithFilePath(t *testing.T) {
                {path.Join("sub", "foobar.html"), "sub", L("sub/single.html", "_default/single.html")},
                {path.Join("content", "foobar.html"), "", L("page/single.html", "_default/single.html")},
                {path.Join("content", "sub", "foobar.html"), "sub", L("sub/single.html", "_default/single.html")},
-               {path.Join("content", "dub", "sub", "foobar.html"), "dub/sub", L("dub/sub/single.html", "dub/single.html", "_default/single.html")},
+               {path.Join("content", "dub", "sub", "foobar.html"), "dub", L("dub/single.html", "_default/single.html")},
        }
 
        for _, el := range toCheck {
                p, err := NewPageFrom(strings.NewReader(SIMPLE_PAGE_YAML), el.input)
-               p.guessSection()
                if err != nil {
                        t.Errorf("Reading from SIMPLE_PAGE_YAML resulted in an error: %s", err)
                }
-               if p.Section != el.section {
-                       t.Errorf("Section not set to %s for page %s. Got: %s", el.section, el.input, p.Section)
+               if p.Section() != el.section {
+                       t.Errorf("Section not set to %s for page %s. Got: %s", el.section, el.input, p.Section())
                }
 
                for _, y := range el.layout {
index 24e3334bbde9ac6a6beaf1af0bc6e7b07dcf03f7..323128041e394d3c11216730c2310dd2f16e4bed 100644 (file)
@@ -3,7 +3,6 @@ package hugolib
 import (
        "errors"
        "fmt"
-       "path/filepath"
        "strconv"
        "strings"
 
@@ -120,9 +119,9 @@ func pageToPermalinkTitle(p *Page, _ string) (string, error) {
 
 // pageToPermalinkFilename returns the URL-safe form of the filename
 func pageToPermalinkFilename(p *Page, _ string) (string, error) {
-       var extension = filepath.Ext(p.FileName)
-       var name = p.FileName[0 : len(p.FileName)-len(extension)]
-       return helpers.Urlize(name), nil
+       //var extension = p.Source.Ext
+       //var name = p.Source.Path()[0 : len(p.Source.Path())-len(extension)]
+       return helpers.Urlize(p.Source.BaseFileName()), nil
 }
 
 // if the page has a slug, return the slug, else return the title
@@ -143,7 +142,7 @@ func pageToPermalinkSlugElseTitle(p *Page, a string) (string, error) {
 
 func pageToPermalinkSection(p *Page, _ string) (string, error) {
        // Page contains Node contains UrlPath which has Section
-       return p.Section, nil
+       return p.Section(), nil
 }
 
 func init() {
index 246312746d350250f8e3e070c5de48cceab2ab94..ae32f9a38a2557a527a383dbb6933799acc53933 100644 (file)
@@ -11,7 +11,7 @@ func (s *Site) ShowPlan(out io.Writer) (err error) {
        }
 
        for _, p := range s.Pages {
-               fmt.Fprintf(out, "%s", p.FileName)
+               fmt.Fprintf(out, "%s", p.Source.Path())
                if p.IsRenderable() {
                        fmt.Fprintf(out, " (renderer: markdown)")
                } else {
index 1635ece994c5f1e49e78d0e369d7065fa7276c7a..ef413bfb30bad52730abe3b9aeacfcc885f1d126 100644 (file)
@@ -20,6 +20,7 @@ import (
        "strings"
        "unicode"
 
+       "github.com/spf13/hugo/helpers"
        jww "github.com/spf13/jwalterweatherman"
 )
 
@@ -93,7 +94,7 @@ func ShortcodesHandle(stringToParse string, p *Page, t Template) string {
                        var data = &ShortcodeWithPage{Params: params, Page: p}
                        if endStart > 0 {
                                s := stringToParse[leadEnd+3 : leadEnd+endStart]
-                               data.Inner = template.HTML(renderBytes([]byte(CleanP(ShortcodesHandle(s, p, t))), p.guessMarkupType(), p.UniqueId()))
+                               data.Inner = template.HTML(helpers.RenderBytes([]byte(CleanP(ShortcodesHandle(s, p, t))), p.guessMarkupType(), p.UniqueId()))
                                remainder := CleanP(stringToParse[leadEnd+endEnd:])
 
                                return CleanP(stringToParse[:leadStart]) +
index ad7afd725dd4a2b8ca729fbcd76d00ccdb072562..d24036eb3bb852361459081c0f16a4e115f954cb 100644 (file)
@@ -236,7 +236,7 @@ func (s *Site) Render() (err error) {
 func (s *Site) checkDescriptions() {
        for _, p := range s.Pages {
                if len(p.Description) < 60 {
-                       jww.FEEDBACK.Println(p.FileName + " ")
+                       jww.FEEDBACK.Println(p.Source.Path() + " ")
                }
        }
 }
@@ -340,7 +340,7 @@ func (s *Site) CreatePages() error {
 
        for i := 0; i < procs*4; i++ {
                wg.Add(1)
-               go pageReader(s, filechan, results, wg)
+               go sourceReader(s, filechan, results, wg)
        }
 
        errs := make(chan error)
@@ -397,18 +397,16 @@ func (s *Site) CreatePages() error {
        return fmt.Errorf("%s\n%s", readErrs, renderErrs)
 }
 
-func pageReader(s *Site, files <-chan *source.File, results chan<- pageResult, wg *sync.WaitGroup) {
+func sourceReader(s *Site, files <-chan *source.File, results chan<- pageResult, wg *sync.WaitGroup) {
        defer wg.Done()
        for file := range files {
-               page, err := NewPage(file.LogicalName)
+               page, err := NewPage(file.Path())
                if err != nil {
                        results <- pageResult{nil, err}
                        continue
                }
                page.Site = &s.Info
                page.Tmpl = s.Tmpl
-               page.Section = file.Section
-               page.Dir = file.Dir
                if err := page.ReadFrom(file.Contents); err != nil {
                        results <- pageResult{nil, err}
                        continue
@@ -604,7 +602,7 @@ func (s *Site) assembleTaxonomies() {
                                        x := WeightedPage{weight.(int), p}
                                        s.Taxonomies[plural].Add(v, x)
                                } else {
-                                       jww.ERROR.Printf("Invalid %s in %s\n", plural, p.File.FileName)
+                                       jww.ERROR.Printf("Invalid %s in %s\n", plural, p.File.Path())
                                }
                        }
                }
@@ -620,7 +618,7 @@ func (s *Site) assembleTaxonomies() {
 
 func (s *Site) assembleSections() {
        for i, p := range s.Pages {
-               s.Sections.Add(p.Section, WeightedPage{s.Pages[i].Weight, s.Pages[i]})
+               s.Sections.Add(p.Section(), WeightedPage{s.Pages[i].Weight, s.Pages[i]})
        }
 
        for k := range s.Sections {
index d4382184834e9a35859c4c500b41725593c49c12..801738ca304ea67378e1a02561790536a7bdb749 100644 (file)
@@ -47,13 +47,13 @@ func checkShowPlanExpected(t *testing.T, s *Site, expected string) {
 
        for _, x := range gotList {
                if !stringInSlice(x, expectedList) {
-                       t.Errorf("ShowPlan expected:\n%q\ngot\n%q", expected, got)
+                       t.Errorf("\nShowPlan expected:\n%q\ngot:\n%q", expected, got)
                }
        }
 
        for _, x := range expectedList {
                if !stringInSlice(x, gotList) {
-                       t.Errorf("ShowPlan expected:\n%q\ngot\n%q", expected, got)
+                       t.Errorf("\nShowPlan expected:\n%q\ngot:\n%q", expected, got)
                }
        }
 }
index a858bf3bb08c5f76b847d7cdde23d420042ec531..044cd062bcadded9ef36567be953e722313bf9ef 100644 (file)
@@ -213,7 +213,7 @@ func TestTargetPath(t *testing.T) {
                        t.Errorf("%s => OutFile  expected: '%s', got: '%s'", test.doc, expected, p.TargetPath())
                }
 
-               if p.Section != test.expectedSection {
+               if p.Section() != test.expectedSection {
                        t.Errorf("%s => p.Section expected: %s, got: %s", test.doc, test.expectedSection, p.Section)
                }
        }
@@ -223,10 +223,10 @@ func TestDraftAndFutureRender(t *testing.T) {
        files := make(map[string][]byte)
        target := &target.InMemoryTarget{Files: files}
        sources := []source.ByteSource{
-               {"sect/doc1.md", []byte("---\ntitle: doc1\ndraft: true\npublishdate: \"2414-05-29\"\n---\n# doc1\n*some content*"), "sect"},
-               {"sect/doc2.md", []byte("---\ntitle: doc2\ndraft: true\npublishdate: \"2012-05-29\"\n---\n# doc2\n*some content*"), "sect"},
-               {"sect/doc3.md", []byte("---\ntitle: doc3\ndraft: false\npublishdate: \"2414-05-29\"\n---\n# doc3\n*some content*"), "sect"},
-               {"sect/doc4.md", []byte("---\ntitle: doc4\ndraft: false\npublishdate: \"2012-05-29\"\n---\n# doc4\n*some content*"), "sect"},
+               {"sect/doc1.md", []byte("---\ntitle: doc1\ndraft: true\npublishdate: \"2414-05-29\"\n---\n# doc1\n*some content*")},
+               {"sect/doc2.md", []byte("---\ntitle: doc2\ndraft: true\npublishdate: \"2012-05-29\"\n---\n# doc2\n*some content*")},
+               {"sect/doc3.md", []byte("---\ntitle: doc3\ndraft: false\npublishdate: \"2414-05-29\"\n---\n# doc3\n*some content*")},
+               {"sect/doc4.md", []byte("---\ntitle: doc4\ndraft: false\npublishdate: \"2012-05-29\"\n---\n# doc4\n*some content*")},
        }
 
        siteSetup := func() *Site {
@@ -283,14 +283,14 @@ func TestSkipRender(t *testing.T) {
        files := make(map[string][]byte)
        target := &target.InMemoryTarget{Files: files}
        sources := []source.ByteSource{
-               {"sect/doc1.html", []byte("---\nmarkup: markdown\n---\n# title\nsome *content*"), "sect"},
-               {"sect/doc2.html", []byte("<!doctype html><html><body>more content</body></html>"), "sect"},
-               {"sect/doc3.md", []byte("# doc3\n*some* content"), "sect"},
-               {"sect/doc4.md", []byte("---\ntitle: doc4\n---\n# doc4\n*some content*"), "sect"},
-               {"sect/doc5.html", []byte("<!doctype html><html>{{ template \"head\" }}<body>body5</body></html>"), "sect"},
-               {"sect/doc6.html", []byte("<!doctype html><html>{{ template \"head_abs\" }}<body>body5</body></html>"), "sect"},
-               {"doc7.html", []byte("<html><body>doc7 content</body></html>"), ""},
-               {"sect/doc8.html", []byte("---\nmarkup: md\n---\n# title\nsome *content*"), "sect"},
+               {"sect/doc1.html", []byte("---\nmarkup: markdown\n---\n# title\nsome *content*")},
+               {"sect/doc2.html", []byte("<!doctype html><html><body>more content</body></html>")},
+               {"sect/doc3.md", []byte("# doc3\n*some* content")},
+               {"sect/doc4.md", []byte("---\ntitle: doc4\n---\n# doc4\n*some content*")},
+               {"sect/doc5.html", []byte("<!doctype html><html>{{ template \"head\" }}<body>body5</body></html>")},
+               {"sect/doc6.html", []byte("<!doctype html><html>{{ template \"head_abs\" }}<body>body5</body></html>")},
+               {"doc7.html", []byte("<html><body>doc7 content</body></html>")},
+               {"sect/doc8.html", []byte("---\nmarkup: md\n---\n# title\nsome *content*")},
        }
 
        viper.Set("verbose", true)
@@ -350,8 +350,8 @@ func TestAbsUrlify(t *testing.T) {
        files := make(map[string][]byte)
        target := &target.InMemoryTarget{Files: files}
        sources := []source.ByteSource{
-               {"sect/doc1.html", []byte("<!doctype html><html><head></head><body><a href=\"#frag1\">link</a></body></html>"), "sect"},
-               {"content/blue/doc2.html", []byte("---\nf: t\n---\n<!doctype html><html><body>more content</body></html>"), "blue"},
+               {"sect/doc1.html", []byte("<!doctype html><html><head></head><body><a href=\"#frag1\">link</a></body></html>")},
+               {"content/blue/doc2.html", []byte("---\nf: t\n---\n<!doctype html><html><body>more content</body></html>")},
        }
        for _, canonify := range []bool{true, false} {
                viper.Set("CanonifyUrls", canonify)
@@ -428,10 +428,10 @@ date = "2012-01-01"
 Front Matter with Ordered Pages 4. This is longer content`)
 
 var WEIGHTED_SOURCES = []source.ByteSource{
-       {"sect/doc1.md", WEIGHTED_PAGE_1, "sect"},
-       {"sect/doc2.md", WEIGHTED_PAGE_2, "sect"},
-       {"sect/doc3.md", WEIGHTED_PAGE_3, "sect"},
-       {"sect/doc4.md", WEIGHTED_PAGE_4, "sect"},
+       {"sect/doc1.md", WEIGHTED_PAGE_1},
+       {"sect/doc2.md", WEIGHTED_PAGE_2},
+       {"sect/doc3.md", WEIGHTED_PAGE_3},
+       {"sect/doc4.md", WEIGHTED_PAGE_4},
 }
 
 func TestOrderedPages(t *testing.T) {
@@ -484,13 +484,18 @@ func TestOrderedPages(t *testing.T) {
 }
 
 var GROUPED_SOURCES = []source.ByteSource{
-       {"sect1/doc1.md", WEIGHTED_PAGE_1, "sect1"},
-       {"sect1/doc2.md", WEIGHTED_PAGE_2, "sect1"},
-       {"sect2/doc3.md", WEIGHTED_PAGE_3, "sect2"},
-       {"sect3/doc4.md", WEIGHTED_PAGE_4, "sect3"},
+       {"sect1/doc1.md", WEIGHTED_PAGE_1},
+       {"sect1/doc2.md", WEIGHTED_PAGE_2},
+       {"sect2/doc3.md", WEIGHTED_PAGE_3},
+       {"sect3/doc4.md", WEIGHTED_PAGE_4},
 }
 
 func TestGroupedPages(t *testing.T) {
+       defer func() {
+               if r := recover(); r != nil {
+                       fmt.Println("Recovered in f", r)
+               }
+       }()
        files := make(map[string][]byte)
        target := &target.InMemoryTarget{Files: files}
 
@@ -583,9 +588,9 @@ func TestWeightedTaxonomies(t *testing.T) {
        files := make(map[string][]byte)
        target := &target.InMemoryTarget{Files: files}
        sources := []source.ByteSource{
-               {"sect/doc1.md", PAGE_WITH_WEIGHTED_TAXONOMIES_1, "sect"},
-               {"sect/doc2.md", PAGE_WITH_WEIGHTED_TAXONOMIES_2, "sect"},
-               {"sect/doc3.md", PAGE_WITH_WEIGHTED_TAXONOMIES_3, "sect"},
+               {"sect/doc1.md", PAGE_WITH_WEIGHTED_TAXONOMIES_1},
+               {"sect/doc2.md", PAGE_WITH_WEIGHTED_TAXONOMIES_2},
+               {"sect/doc3.md", PAGE_WITH_WEIGHTED_TAXONOMIES_3},
        }
        taxonomies := make(map[string]string)
 
index 63b737c316e59150114e951040c25674932d62ff..e82a8e4c5c383d70b0b6ba9098b5d71f83be7db7 100644 (file)
@@ -45,8 +45,8 @@ func (t *InMemoryAliasTarget) Publish(label string, permalink template.HTML) (er
 }
 
 var urlFakeSource = []source.ByteSource{
-       {"content/blue/doc1.md", []byte(SLUG_DOC_1), "blue"},
-       {"content/blue/doc2.md", []byte(SLUG_DOC_2), "blue"},
+       {"content/blue/doc1.md", []byte(SLUG_DOC_1)},
+       {"content/blue/doc2.md", []byte(SLUG_DOC_2)},
 }
 
 func TestPageCount(t *testing.T) {
index 3b1aa6b39ae7a3866de637785c81012b82548806..938c77a2358efca32651505a71b9d9af8163544e 100644 (file)
@@ -1,76 +1 @@
 package hugolib
-
-import (
-       "bytes"
-       "os/exec"
-       "strings"
-
-       jww "github.com/spf13/jwalterweatherman"
-)
-
-var summaryLength = 70
-var summaryDivider = []byte("<!--more-->")
-
-func TotalWords(s string) int {
-       return len(strings.Fields(s))
-}
-
-func WordCount(s string) map[string]int {
-       m := make(map[string]int)
-       for _, f := range strings.Fields(s) {
-               m[f] += 1
-       }
-
-       return m
-}
-
-func RemoveSummaryDivider(content []byte) []byte {
-       return bytes.Replace(content, summaryDivider, []byte(""), -1)
-}
-
-func TruncateWords(s string, max int) string {
-       words := strings.Fields(s)
-       if max > len(words) {
-               return strings.Join(words, " ")
-       }
-
-       return strings.Join(words[:max], " ")
-}
-
-func TruncateWordsToWholeSentence(s string, max int) string {
-       words := strings.Fields(s)
-       if max > len(words) {
-               return strings.Join(words, " ")
-       }
-
-       for counter, word := range words[max:] {
-               if strings.HasSuffix(word, ".") ||
-                       strings.HasSuffix(word, "?") ||
-                       strings.HasSuffix(word, ".\"") ||
-                       strings.HasSuffix(word, "!") {
-                       return strings.Join(words[:max+counter+1], " ")
-               }
-       }
-
-       return strings.Join(words[:max], " ")
-}
-
-func getRstContent(content []byte) string {
-       cleanContent := bytes.Replace(content, summaryDivider, []byte(""), 1)
-
-       cmd := exec.Command("rst2html.py", "--leave-comments")
-       cmd.Stdin = bytes.NewReader(cleanContent)
-       var out bytes.Buffer
-       cmd.Stdout = &out
-       if err := cmd.Run(); err != nil {
-               jww.ERROR.Println(err)
-       }
-
-       rstLines := strings.Split(out.String(), "\n")
-       for i, line := range rstLines {
-               if strings.HasPrefix(line, "<body>") {
-                       rstLines = (rstLines[i+1 : len(rstLines)-3])
-               }
-       }
-       return strings.Join(rstLines, "\n")
-}
diff --git a/source/file.go b/source/file.go
new file mode 100644 (file)
index 0000000..ce1469d
--- /dev/null
@@ -0,0 +1,114 @@
+// Copyright © 2014 Steve Francia <spf@spf13.com>.
+//
+// Licensed under the Simple Public License, Version 2.0 (the "License");
+// you may not use this file except in compliance with the License.
+// You may obtain a copy of the License at
+// http://opensource.org/licenses/Simple-2.0
+//
+// Unless required by applicable law or agreed to in writing, software
+// distributed under the License is distributed on an "AS IS" BASIS,
+// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+// See the License for the specific language governing permissions and
+// limitations under the License.
+
+package source
+
+import (
+       "io"
+       "path"
+       "path/filepath"
+       "strings"
+
+       "github.com/spf13/hugo/helpers"
+)
+
+type File struct {
+       relpath     string // Original Full Path eg. /Users/Home/Hugo/foo.txt
+       logicalName string // foo.txt
+       Contents    io.Reader
+       section     string // The first directory
+       dir         string // The full directory Path (minus file name)
+       ext         string // Just the ext (eg txt)
+       uniqueId    string // MD5 of the filename
+}
+
+func (f *File) UniqueId() string {
+       if f.uniqueId == "" {
+               f.uniqueId = helpers.Md5String(f.LogicalName())
+       }
+       return f.uniqueId
+}
+
+// Filename without extension
+func (f *File) BaseFileName() string {
+       return helpers.Filename(f.LogicalName())
+}
+
+func (f *File) Section() string {
+       if f.section != "" {
+               return f.section
+       } else {
+               f.section = helpers.GuessSection(f.Dir())
+               return f.section
+       }
+}
+
+func (f *File) LogicalName() string {
+       if f.logicalName != "" {
+               return f.logicalName
+       } else {
+               _, f.logicalName = path.Split(f.relpath)
+               return f.logicalName
+       }
+}
+
+//func (f *File) SetDir(dir string) {
+//f.dir = dir
+//}
+
+func (f *File) Dir() string {
+       if f.dir != "" {
+               return f.dir
+       } else {
+               f.dir, _ = path.Split(f.relpath)
+               return f.dir
+       }
+}
+
+func (f *File) Extension() string {
+       if f.ext != "" {
+               return f.ext
+       } else {
+               f.ext = strings.TrimPrefix(filepath.Ext(f.LogicalName()), ".")
+               return f.ext
+       }
+}
+
+func (f *File) Ext() string {
+       return f.Extension()
+}
+
+func (f *File) Path() string {
+       return f.relpath
+}
+
+func NewFileWithContents(relpath string, content io.Reader) *File {
+       file := NewFile(relpath)
+       file.Contents = content
+       return file
+}
+
+func NewFile(relpath string) *File {
+       return &File{
+               relpath: relpath,
+       }
+}
+
+func NewFileFromAbs(base, fullpath string, content io.Reader) (f *File, err error) {
+       var name string
+       if name, err = helpers.GetRelativePath(fullpath, base); err != nil {
+               return nil, err
+       }
+
+       return NewFileWithContents(name, content), nil
+}
index 3a176d7ba59ef7b69f2f788e4634e011967a6fe3..d89149dc6c000469e491661e616d1bb82030301e 100644 (file)
@@ -1,34 +1,56 @@
+// Copyright © 2014 Steve Francia <spf@spf13.com>.
+//
+// Licensed under the Simple Public License, Version 2.0 (the "License");
+// you may not use this file except in compliance with the License.
+// You may obtain a copy of the License at
+// http://opensource.org/licenses/Simple-2.0
+//
+// Unless required by applicable law or agreed to in writing, software
+// distributed under the License is distributed on an "AS IS" BASIS,
+// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+// See the License for the specific language governing permissions and
+// limitations under the License.
+
 package source
 
 import (
        "bytes"
-       "errors"
        "io"
        "io/ioutil"
        "os"
-       "path"
        "path/filepath"
        "strings"
+
+       "github.com/spf13/hugo/helpers"
 )
 
 type Input interface {
        Files() []*File
 }
 
-type File struct {
-       name        string
-       LogicalName string
-       Contents    io.Reader
-       Section     string
-       Dir         string
-}
-
 type Filesystem struct {
        files      []*File
        Base       string
        AvoidPaths []string
 }
 
+func (f *Filesystem) FilesByExts(exts ...string) []*File {
+       var newFiles []*File
+
+       if len(exts) == 0 {
+               return f.Files()
+       }
+
+       for _, x := range f.Files() {
+               for _, e := range exts {
+                       if x.Ext() == strings.TrimPrefix(e, ".") {
+                               newFiles = append(newFiles, x)
+                       }
+               }
+       }
+       return newFiles
+}
+
 func (f *Filesystem) Files() []*File {
        if len(f.files) < 1 {
                f.captureFiles()
@@ -36,47 +58,23 @@ func (f *Filesystem) Files() []*File {
        return f.files
 }
 
-var errMissingBaseDir = errors.New("source: missing base directory")
-
 func (f *Filesystem) add(name string, reader io.Reader) (err error) {
+       var file *File
 
-       if name, err = f.getRelativePath(name); err != nil {
-               return err
-       }
-
-       // section should be the first part of the path
-       dir, logical := path.Split(name)
-       parts := strings.Split(dir, "/")
-       section := parts[0]
+       //if f.Base == "" {
+       //file = NewFileWithContents(name, reader)
+       //} else {
+       file, err = NewFileFromAbs(f.Base, name, reader)
+       //}
 
-       if section == "." {
-               section = ""
+       if err == nil {
+               f.files = append(f.files, file)
        }
-
-       f.files = append(f.files, &File{
-               name:        name,
-               LogicalName: logical,
-               Contents:    reader,
-               Section:     section,
-               Dir:         dir,
-       })
-
-       return
+       return err
 }
 
 func (f *Filesystem) getRelativePath(name string) (final string, err error) {
-       if filepath.IsAbs(name) && f.Base == "" {
-               return "", errMissingBaseDir
-       }
-       name = filepath.Clean(name)
-       base := filepath.Clean(f.Base)
-
-       name, err = filepath.Rel(base, name)
-       if err != nil {
-               return "", err
-       }
-       name = filepath.ToSlash(name)
-       return name, nil
+       return helpers.GetRelativePath(name, f.Base)
 }
 
 func (f *Filesystem) captureFiles() {
index b885d2f24611790f03edc4b18fa0015938c605e8..d2639310e8d99c055c0d5ef045c792b4a15ed05f 100644 (file)
@@ -32,16 +32,17 @@ func TestAddFile(t *testing.T) {
                }
 
                for _, src := range []*Filesystem{srcDefault, srcWithBase} {
+
                        p := test.filename
                        if !filepath.IsAbs(test.filename) {
                                p = path.Join(src.Base, test.filename)
                        }
 
                        if err := src.add(p, bytes.NewReader([]byte(test.content))); err != nil {
-                               if err == errMissingBaseDir {
+                               if err.Error() == "source: missing base directory" {
                                        continue
                                }
-                               t.Fatalf("%s add returned and error: %s", p, err)
+                               t.Fatalf("%s add returned an error: %s", p, err)
                        }
 
                        if len(src.Files()) != 1 {
@@ -49,8 +50,8 @@ func TestAddFile(t *testing.T) {
                        }
 
                        f := src.Files()[0]
-                       if f.LogicalName != test.logical {
-                               t.Errorf("Filename (Base: %q) expected: %q, got: %q", src.Base, test.logical, f.LogicalName)
+                       if f.LogicalName() != test.logical {
+                               t.Errorf("Filename (Base: %q) expected: %q, got: %q", src.Base, test.logical, f.LogicalName())
                        }
 
                        b := new(bytes.Buffer)
@@ -59,12 +60,12 @@ func TestAddFile(t *testing.T) {
                                t.Errorf("File (Base: %q) contents should be %q, got: %q", src.Base, test.content, b.String())
                        }
 
-                       if f.Section != test.section {
-                               t.Errorf("File section (Base: %q) expected: %q, got: %q", src.Base, test.section, f.Section)
+                       if f.Section() != test.section {
+                               t.Errorf("File section (Base: %q) expected: %q, got: %q", src.Base, test.section, f.Section())
                        }
 
-                       if f.Dir != test.dir {
-                               t.Errorf("Dir path (Base: %q) expected: %q, got: %q", src.Base, test.dir, f.Dir)
+                       if f.Dir() != test.dir {
+                               t.Errorf("Dir path (Base: %q) expected: %q, got: %q", src.Base, test.dir, f.Dir())
                        }
                }
        }
index 7c76469fec64b77ff5ceeb035970eea09b3e576d..c30aa38360b00e661218894e6ad5a46a12af316e 100644 (file)
@@ -3,17 +3,15 @@ package source
 import (
        "bytes"
        "fmt"
-       "path"
 )
 
 type ByteSource struct {
        Name    string
        Content []byte
-       Section string
 }
 
 func (b *ByteSource) String() string {
-       return fmt.Sprintf("%s %s %s", b.Name, b.Section, string(b.Content))
+       return fmt.Sprintf("%s %s", b.Name, string(b.Content))
 }
 
 type InMemorySource struct {
@@ -23,12 +21,7 @@ type InMemorySource struct {
 func (i *InMemorySource) Files() (files []*File) {
        files = make([]*File, len(i.ByteSource))
        for i, fake := range i.ByteSource {
-               files[i] = &File{
-                       LogicalName: fake.Name,
-                       Contents:    bytes.NewReader(fake.Content),
-                       Section:     fake.Section,
-                       Dir:         path.Dir(fake.Name),
-               }
+               files[i] = NewFileWithContents(fake.Name, bytes.NewReader(fake.Content))
        }
        return
 }