blob: 2c5b48d29f4ef7ab6516f2f7a54e3897aa5e9326 [file]
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
########################################################################
# OPENAPI-URI: /api/code/pony
########################################################################
# get:
# responses:
# '200':
# content:
# application/json:
# schema:
# $ref: '#/components/schemas/Factor'
# description: 200 Response
# default:
# content:
# application/json:
# schema:
# $ref: '#/components/schemas/Error'
# description: unexpected error
# security:
# - cookieAuth: []
# summary: Shows pony factor data for a set of repos over a given period of time
# post:
# requestBody:
# content:
# application/json:
# schema:
# $ref: '#/components/schemas/defaultWidgetArgs'
# responses:
# '200':
# content:
# application/json:
# schema:
# $ref: '#/components/schemas/Factor'
# description: 200 Response
# default:
# content:
# application/json:
# schema:
# $ref: '#/components/schemas/Error'
# description: unexpected error
# security:
# - cookieAuth: []
# summary: Shows pony factor data for a set of repos over a given period of time
#
########################################################################
"""
This is the pony factor renderer for Kibble
"""
import json
import time
import re
def run(API, environ, indata, session):
# We need to be logged in for this!
if not session.user:
raise API.exception(403, "You must be logged in to use this API endpoint! %s")
now = time.time()
# First, fetch the view if we have such a thing enabled
viewList = []
if indata.get("view"):
viewList = session.getView(indata.get("view"))
if indata.get("subfilter"):
viewList = session.subFilter(indata.get("subfilter"), view=viewList)
dateTo = indata.get("to", int(time.time()))
dateFrom = indata.get(
"from", dateTo - (86400 * 30 * 24)
) # Default to a 24 month span
if dateFrom < 0:
dateFrom = 0
dateYonder = dateFrom - (dateTo - dateFrom)
####################################################################
####################################################################
dOrg = session.user["defaultOrganisation"] or "apache"
query = {
"query": {
"bool": {
"must": [
{"range": {"tsday": {"from": dateFrom, "to": dateTo}}},
{"term": {"organisation": dOrg}},
]
}
}
}
# Source-specific or view-specific??
if indata.get("source"):
query["query"]["bool"]["must"].append(
{"term": {"sourceID": indata.get("source")}}
)
elif viewList:
query["query"]["bool"]["must"].append({"terms": {"sourceID": viewList}})
# Get an initial count of commits
res = session.DB.ES.count(
index=session.DB.dbname, doc_type="code_commit", body=query
)
globcount = res["count"]
# Get top 25 committers this period
query["aggs"] = {
"by_committer": {"terms": {"field": "committer_email", "size": 5000}},
"by_author": {"terms": {"field": "author_email", "size": 5000}},
}
res = session.DB.ES.search(
index=session.DB.dbname, doc_type="code_commit", size=0, body=query
)
# PF for committers
pf_committer = 0
pf_committer_count = 0
for bucket in res["aggregations"]["by_committer"]["buckets"]:
count = bucket["doc_count"]
pf_committer += 1
pf_committer_count += count
if pf_committer_count > int(globcount / 2):
break
# PF for authors
pf_author = 0
pf_author_count = 0
cpf = {}
for bucket in res["aggregations"]["by_author"]["buckets"]:
count = bucket["doc_count"]
pf_author += 1
pf_author_count += count
mldom = bucket["key"].lower().split("@")[1]
cpf[mldom] = True
if pf_author_count > int(globcount / 2):
break
####################################################################
####################################################################
dOrg = session.user["defaultOrganisation"] or "apache"
query = {
"query": {
"bool": {
"must": [
{"range": {"tsday": {"from": dateYonder, "to": dateFrom - 1}}},
{"term": {"organisation": dOrg}},
]
}
}
}
# Source-specific or view-specific??
if indata.get("source"):
query["query"]["bool"]["must"].append(
{"term": {"sourceID": indata.get("source")}}
)
elif viewList:
query["query"]["bool"]["must"].append({"terms": {"sourceID": viewList}})
# Get an initial count of commits
res = session.DB.ES.count(
index=session.DB.dbname, doc_type="code_commit", body=query
)
globcount = res["count"]
# Get top 25 committers this period
query["aggs"] = {
"by_committer": {"terms": {"field": "committer_email", "size": 5000}},
"by_author": {"terms": {"field": "author_email", "size": 5000}},
}
res = session.DB.ES.search(
index=session.DB.dbname, doc_type="code_commit", size=0, body=query
)
# PF for committers
pf_committer_b = 0
pf_committer_count = 0
for bucket in res["aggregations"]["by_committer"]["buckets"]:
count = bucket["doc_count"]
pf_committer_b += 1
pf_committer_count += count
if pf_committer_count > int(globcount / 2):
break
# PF for authors
pf_author_b = 0
pf_author_count = 0
cpf_b = {}
for bucket in res["aggregations"]["by_author"]["buckets"]:
count = bucket["doc_count"]
pf_author_b += 1
pf_author_count += count
mldom = bucket["key"].lower().split("@")[1]
cpf_b[mldom] = True
if pf_author_count > int(globcount / 2):
break
JSON_OUT = {
"factors": [
{
"title": "Pony Factor (by committership)",
"count": pf_committer,
"previous": pf_committer_b,
},
{
"title": "Pony Factor (by authorship)",
"count": pf_author,
"previous": pf_author_b,
},
{
"title": "Meta-Pony Factor (by authorship)",
"count": len(cpf),
"previous": len(cpf_b),
},
],
"okay": True,
"responseTime": time.time() - now,
}
yield json.dumps(JSON_OUT)