blob: 4bc0d36d4498a4e07ac863602966e3e92f08f147 [file]
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
########################################################################
# OPENAPI-URI: /api/code/committers
########################################################################
# get:
# responses:
# '200':
# content:
# application/json:
# schema:
# $ref: '#/components/schemas/CommitterList'
# description: 200 Response
# default:
# content:
# application/json:
# schema:
# $ref: '#/components/schemas/Error'
# description: unexpected error
# security:
# - cookieAuth: []
# summary: Shows the top N of committers
# post:
# requestBody:
# content:
# application/json:
# schema:
# $ref: '#/components/schemas/defaultWidgetArgs'
# responses:
# '200':
# content:
# application/json:
# schema:
# $ref: '#/components/schemas/CommitterList'
# description: 200 Response
# default:
# content:
# application/json:
# schema:
# $ref: '#/components/schemas/Error'
# description: unexpected error
# security:
# - cookieAuth: []
# summary: Shows trend data for a set of repos over a given period of time
#
########################################################################
"""
This is the TopN committers list renderer for Kibble
"""
import json
import time
import hashlib
def run(API, environ, indata, session):
# We need to be logged in for this!
if not session.user:
raise API.exception(403, "You must be logged in to use this API endpoint! %s")
now = time.time()
# First, fetch the view if we have such a thing enabled
viewList = []
if indata.get("view"):
viewList = session.getView(indata.get("view"))
if indata.get("subfilter"):
viewList = session.subFilter(indata.get("subfilter"), view=viewList)
dateTo = indata.get("to", int(time.time()))
dateFrom = indata.get(
"from", dateTo - (86400 * 30 * 6)
) # Default to a 6 month span
which = "committer_email"
role = "committer"
if indata.get("author", False):
which = "author_email"
role = "author"
interval = indata.get("interval", "month")
####################################################################
####################################################################
dOrg = session.user["defaultOrganisation"] or "apache"
query = {
"query": {
"bool": {
"must": [
{"range": {"tsday": {"from": dateFrom, "to": dateTo}}},
{"term": {"organisation": dOrg}},
]
}
}
}
# Source-specific or view-specific??
if indata.get("source"):
query["query"]["bool"]["must"].append(
{"term": {"sourceID": indata.get("source")}}
)
elif viewList:
query["query"]["bool"]["must"].append({"terms": {"sourceID": viewList}})
if indata.get("email"):
query["query"]["bool"]["should"] = [
{"term": {"committer_email": indata.get("email")}},
{"term": {"author_email": indata.get("email")}},
]
query["query"]["bool"]["minimum_should_match"] = 1
# Path filter?
if indata.get("pathfilter"):
pf = indata.get("pathfilter")
if "!" in pf:
pf = pf.replace("!", "")
query["query"]["bool"]["must_not"] = query["query"]["bool"].get(
"must_not", []
)
query["query"]["bool"]["must_not"].append({"regexp": {"files_changed": pf}})
else:
query["query"]["bool"]["must"].append({"regexp": {"files_changed": pf}})
# Get top 25 committers this period
query["aggs"] = {
"committers": {
"terms": {"field": which, "size": 25},
"aggs": {
"byinsertions": {
"terms": {"field": which},
"aggs": {"stats": {"sum": {"field": "insertions"}}},
},
"bydeletions": {
"terms": {"field": which},
"aggs": {"stats": {"sum": {"field": "deletions"}}},
},
},
}
}
res = session.DB.ES.search(
index=session.DB.dbname, doc_type="code_commit", size=0, body=query
)
people = {}
for bucket in res["aggregations"]["committers"]["buckets"]:
email = bucket["key"]
count = bucket["doc_count"]
sha = hashlib.sha1(("%s%s" % (dOrg, email)).encode("utf-8")).hexdigest()
if session.DB.ES.exists(index=session.DB.dbname, doc_type="person", id=sha):
pres = session.DB.ES.get(index=session.DB.dbname, doc_type="person", id=sha)
person = pres["_source"]
person["name"] = person.get("name", "unknown")
people[email] = person
people[email]["gravatar"] = hashlib.md5(
person.get("email", "unknown").encode("utf-8")
).hexdigest()
people[email]["count"] = count
people[email]["subcount"] = {
"insertions": int(
bucket["byinsertions"]["buckets"][0]["stats"]["value"]
),
"deletions": int(bucket["bydeletions"]["buckets"][0]["stats"]["value"]),
}
topN = []
for email, person in people.items():
topN.append(person)
topN = sorted(topN, key=lambda x: x["count"], reverse=True)
# Get timeseries for this period
query["aggs"] = {
"per_interval": {
"date_histogram": {"field": "date", "interval": interval},
"aggs": {
"by_committer": {"cardinality": {"field": "committer_email"}},
"by_author": {"cardinality": {"field": "author_email"}},
},
}
}
res = session.DB.ES.search(
index=session.DB.dbname, doc_type="code_commit", size=0, body=query
)
timeseries = []
for bucket in res["aggregations"]["per_interval"]["buckets"]:
ts = int(bucket["key"] / 1000)
ccount = bucket["by_committer"]["value"]
acount = bucket["by_author"]["value"]
timeseries.append({"date": ts, "committers": ccount, "authors": acount})
JSON_OUT = {
"topN": {"denoter": "commits", "items": topN},
"timeseries": timeseries,
"sorted": people,
"okay": True,
"responseTime": time.time() - now,
"widgetType": {"chartType": "bar"},
}
yield json.dumps(JSON_OUT)