mirror of
https://github.com/go-gitea/gitea
synced 2024-11-16 15:14:24 +00:00
110 lines
2.7 KiB
Go
110 lines
2.7 KiB
Go
|
// Copyright 2020 The Gitea Authors. All rights reserved.
|
||
|
// Use of this source code is governed by a MIT-style
|
||
|
// license that can be found in the LICENSE file.
|
||
|
|
||
|
// +build !gogit
|
||
|
|
||
|
package git
|
||
|
|
||
|
import (
|
||
|
"bytes"
|
||
|
"io"
|
||
|
"io/ioutil"
|
||
|
|
||
|
"code.gitea.io/gitea/modules/analyze"
|
||
|
|
||
|
"github.com/go-enry/go-enry/v2"
|
||
|
)
|
||
|
|
||
|
// GetLanguageStats calculates language stats for git repository at specified commit
|
||
|
func (repo *Repository) GetLanguageStats(commitID string) (map[string]int64, error) {
|
||
|
// FIXME: We can be more efficient here...
|
||
|
//
|
||
|
// We're expecting that we will be reading a lot of blobs and the trees
|
||
|
// Thus we should use a shared `cat-file --batch` to get all of this data
|
||
|
// And keep the buffers around with resets as necessary.
|
||
|
//
|
||
|
// It's more complicated so...
|
||
|
commit, err := repo.GetCommit(commitID)
|
||
|
if err != nil {
|
||
|
log("Unable to get commit for: %s", commitID)
|
||
|
return nil, err
|
||
|
}
|
||
|
|
||
|
tree := commit.Tree
|
||
|
|
||
|
entries, err := tree.ListEntriesRecursive()
|
||
|
if err != nil {
|
||
|
return nil, err
|
||
|
}
|
||
|
|
||
|
sizes := make(map[string]int64)
|
||
|
for _, f := range entries {
|
||
|
if f.Size() == 0 || enry.IsVendor(f.Name()) || enry.IsDotFile(f.Name()) ||
|
||
|
enry.IsDocumentation(f.Name()) || enry.IsConfiguration(f.Name()) {
|
||
|
continue
|
||
|
}
|
||
|
|
||
|
// If content can not be read or file is too big just do detection by filename
|
||
|
var content []byte
|
||
|
if f.Size() <= bigFileSize {
|
||
|
content, _ = readFile(f, fileSizeLimit)
|
||
|
}
|
||
|
if enry.IsGenerated(f.Name(), content) {
|
||
|
continue
|
||
|
}
|
||
|
|
||
|
// TODO: Use .gitattributes file for linguist overrides
|
||
|
// FIXME: Why can't we split this and the IsGenerated tests to avoid reading the blob unless absolutely necessary?
|
||
|
// - eg. do the all the detection tests using filename first before reading content.
|
||
|
language := analyze.GetCodeLanguage(f.Name(), content)
|
||
|
if language == enry.OtherLanguage || language == "" {
|
||
|
continue
|
||
|
}
|
||
|
|
||
|
// group languages, such as Pug -> HTML; SCSS -> CSS
|
||
|
group := enry.GetLanguageGroup(language)
|
||
|
if group != "" {
|
||
|
language = group
|
||
|
}
|
||
|
|
||
|
sizes[language] += f.Size()
|
||
|
|
||
|
continue
|
||
|
}
|
||
|
|
||
|
// filter special languages unless they are the only language
|
||
|
if len(sizes) > 1 {
|
||
|
for language := range sizes {
|
||
|
langtype := enry.GetLanguageType(language)
|
||
|
if langtype != enry.Programming && langtype != enry.Markup {
|
||
|
delete(sizes, language)
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
return sizes, nil
|
||
|
}
|
||
|
|
||
|
func readFile(entry *TreeEntry, limit int64) ([]byte, error) {
|
||
|
// FIXME: We can probably be a little more efficient here... see above
|
||
|
r, err := entry.Blob().DataAsync()
|
||
|
if err != nil {
|
||
|
return nil, err
|
||
|
}
|
||
|
defer r.Close()
|
||
|
|
||
|
if limit <= 0 {
|
||
|
return ioutil.ReadAll(r)
|
||
|
}
|
||
|
|
||
|
size := entry.Size()
|
||
|
if limit > 0 && size > limit {
|
||
|
size = limit
|
||
|
}
|
||
|
buf := bytes.NewBuffer(nil)
|
||
|
buf.Grow(int(size))
|
||
|
_, err = io.Copy(buf, io.LimitReader(r, limit))
|
||
|
return buf.Bytes(), err
|
||
|
}
|