instance_id
stringlengths
17
39
repo
stringclasses
8 values
issue_id
stringlengths
14
34
pr_id
stringlengths
14
34
linking_methods
sequencelengths
1
3
base_commit
stringlengths
40
40
merge_commit
stringlengths
0
40
βŒ€
hints_text
sequencelengths
0
106
resolved_comments
sequencelengths
0
119
created_at
unknown
labeled_as
sequencelengths
0
7
problem_title
stringlengths
7
174
problem_statement
stringlengths
0
55.4k
gold_files
sequencelengths
0
10
gold_files_postpatch
sequencelengths
1
10
test_files
sequencelengths
0
60
gold_patch
stringlengths
220
5.83M
test_patch
stringlengths
386
194k
βŒ€
split_random
stringclasses
3 values
split_time
stringclasses
3 values
issue_start_time
timestamp[ns]
issue_created_at
unknown
issue_by_user
stringlengths
3
21
split_repo
stringclasses
3 values
provectus/kafka-ui/1572_2640
provectus/kafka-ui
provectus/kafka-ui/1572
provectus/kafka-ui/2640
[ "connected" ]
049b35fc99fd09d270fbccb3f0a2bbcc9bf11f56
97f1c639a3556725ee7205718f6e16f40f3753d3
[ "error occurs when converting recursive avro schema to json schema (stackoverflow in AvroJsonSchemaConverter)\r\nexample failing schema \r\n```\r\n{\r\n\t\"type\": \"record\",\r\n\t\"name\": \"Anything\",\r\n\t\"namespace\": \"io.confluent.connect.avro\",\r\n\t\"fields\": [\r\n\t\t{\r\n\t\t\t\"name\": \"somekey\",\r\n\t\t\t\"type\": \"string\"\r\n\t\t},\r\n\t\t{\r\n\t\t\t\"name\": \"recursion\",\r\n\t\t\t\"type\": \"Anything\"\r\n\t\t}\r\n ]\r\n}\r\n```" ]
[]
"2022-09-26T11:09:25Z"
[ "type/bug", "scope/backend", "status/accepted", "status/confirmed" ]
[Bug] Clicking 'produce message' on some avro-schemed topics breaks
**Describe the bug** <!--(A clear and concise description of what the bug is.)--> Clicking 'produce message' on some avro-schemed topics produces 502. **Set up** docker, latest master <!-- (How do you run the app? Which version of the app are you running? Provide either docker image version or check commit hash at the top left corner. We won't be able to help you without this information.) --> **Steps to Reproduce** Steps to reproduce the behavior: 1. set up reddit connector 2. go to topic reddit-posts-avro 3. click 'produce message' 4. ??? 5. PROFIT **Expected behavior** should work fine <!-- (A clear and concise description of what you expected to happen) --> **Screenshots** <!-- (If applicable, add screenshots to help explain your problem) --> **Additional context** <!-- (Add any other context about the problem here) -->
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/util/jsonschema/AvroJsonSchemaConverter.java" ]
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/util/jsonschema/AvroJsonSchemaConverter.java" ]
[ "kafka-ui-api/src/test/java/com/provectus/kafka/ui/util/jsonschema/AvroJsonSchemaConverterTest.java" ]
diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/util/jsonschema/AvroJsonSchemaConverter.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/util/jsonschema/AvroJsonSchemaConverter.java index 61f742e7214..8c703526139 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/util/jsonschema/AvroJsonSchemaConverter.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/util/jsonschema/AvroJsonSchemaConverter.java @@ -4,7 +4,6 @@ import java.util.Collections; import java.util.HashMap; import java.util.List; -import java.util.Locale; import java.util.Map; import java.util.stream.Collectors; import org.apache.avro.Schema; @@ -22,7 +21,7 @@ public JsonSchema convert(URI basePath, Schema schema) { builder.type(type); Map<String, FieldSchema> definitions = new HashMap<>(); - final FieldSchema root = convertSchema("root", schema, definitions, false); + final FieldSchema root = convertSchema(schema, definitions, true); builder.definitions(definitions); if (type.getType().equals(JsonType.Type.OBJECT)) { @@ -36,11 +35,11 @@ public JsonSchema convert(URI basePath, Schema schema) { private FieldSchema convertField(Schema.Field field, Map<String, FieldSchema> definitions) { - return convertSchema(field.name(), field.schema(), definitions, true); + return convertSchema(field.schema(), definitions, false); } - private FieldSchema convertSchema(String name, Schema schema, - Map<String, FieldSchema> definitions, boolean ref) { + private FieldSchema convertSchema(Schema schema, + Map<String, FieldSchema> definitions, boolean isRoot) { if (!schema.isUnion()) { JsonType type = convertType(schema); switch (type.getType()) { @@ -53,12 +52,12 @@ private FieldSchema convertSchema(String name, Schema schema, return new SimpleFieldSchema(type); case OBJECT: if (schema.getType().equals(Schema.Type.MAP)) { - return new MapFieldSchema(convertSchema(name, schema.getValueType(), definitions, ref)); + return new MapFieldSchema(convertSchema(schema.getValueType(), definitions, isRoot)); } else { - return createObjectSchema(name, schema, definitions, ref); + return createObjectSchema(schema, definitions, isRoot); } case ARRAY: - return createArraySchema(name, schema, definitions); + return createArraySchema(schema, definitions); default: throw new RuntimeException("Unknown type"); } @@ -67,20 +66,26 @@ private FieldSchema convertSchema(String name, Schema schema, } } - private FieldSchema createUnionSchema(Schema schema, Map<String, FieldSchema> definitions) { + // this method formats json-schema field in a way + // to fit avro-> json encoding rules (https://avro.apache.org/docs/1.11.1/specification/_print/#json-encoding) + private FieldSchema createUnionSchema(Schema schema, Map<String, FieldSchema> definitions) { final boolean nullable = schema.getTypes().stream() .anyMatch(t -> t.getType().equals(Schema.Type.NULL)); final Map<String, FieldSchema> fields = schema.getTypes().stream() .filter(t -> !t.getType().equals(Schema.Type.NULL)) - .map(f -> Tuples.of( - f.getType().getName().toLowerCase(Locale.ROOT), - convertSchema( - f.getType().getName().toLowerCase(Locale.ROOT), - f, definitions, true - ) - )).collect(Collectors.toMap( + .map(f -> { + String oneOfFieldName; + if (f.getType().equals(Schema.Type.RECORD)) { + // for records using full record name + oneOfFieldName = f.getFullName(); + } else { + // for primitive types - using type name + oneOfFieldName = f.getType().getName().toLowerCase(); + } + return Tuples.of(oneOfFieldName, convertSchema(f, definitions, false)); + }).collect(Collectors.toMap( Tuple2::getT1, Tuple2::getT2 )); @@ -97,8 +102,16 @@ private FieldSchema createUnionSchema(Schema schema, Map<String, FieldSchema> de } } - private FieldSchema createObjectSchema(String name, Schema schema, - Map<String, FieldSchema> definitions, boolean ref) { + private FieldSchema createObjectSchema(Schema schema, + Map<String, FieldSchema> definitions, + boolean isRoot) { + var definitionName = schema.getFullName(); + if (definitions.containsKey(definitionName)) { + return createRefField(definitionName); + } + // adding stub record, need to avoid infinite recursion + definitions.put(definitionName, new ObjectFieldSchema(Map.of(), List.of())); + final Map<String, FieldSchema> fields = schema.getFields().stream() .map(f -> Tuples.of(f.name(), convertField(f, definitions))) .collect(Collectors.toMap( @@ -110,19 +123,26 @@ private FieldSchema createObjectSchema(String name, Schema schema, .filter(f -> !f.schema().isNullable()) .map(Schema.Field::name).collect(Collectors.toList()); - if (ref) { - String definitionName = String.format("Record%s", schema.getName()); - definitions.put(definitionName, new ObjectFieldSchema(fields, required)); - return new RefFieldSchema(String.format("#/definitions/%s", definitionName)); + var objectSchema = new ObjectFieldSchema(fields, required); + if (isRoot) { + // replacing stub with self-reference (need for usage in json-schema's oneOf) + definitions.put(definitionName, new RefFieldSchema("#")); + return objectSchema; } else { - return new ObjectFieldSchema(fields, required); + // replacing stub record with actual object structure + definitions.put(definitionName, objectSchema); + return createRefField(definitionName); } } - private ArrayFieldSchema createArraySchema(String name, Schema schema, + private RefFieldSchema createRefField(String definitionName) { + return new RefFieldSchema(String.format("#/definitions/%s", definitionName)); + } + + private ArrayFieldSchema createArraySchema(Schema schema, Map<String, FieldSchema> definitions) { return new ArrayFieldSchema( - convertSchema(name, schema.getElementType(), definitions, true) + convertSchema(schema.getElementType(), definitions, false) ); }
diff --git a/kafka-ui-api/src/test/java/com/provectus/kafka/ui/util/jsonschema/AvroJsonSchemaConverterTest.java b/kafka-ui-api/src/test/java/com/provectus/kafka/ui/util/jsonschema/AvroJsonSchemaConverterTest.java index d78426d48a1..24d4daf7d9d 100644 --- a/kafka-ui-api/src/test/java/com/provectus/kafka/ui/util/jsonschema/AvroJsonSchemaConverterTest.java +++ b/kafka-ui-api/src/test/java/com/provectus/kafka/ui/util/jsonschema/AvroJsonSchemaConverterTest.java @@ -1,27 +1,29 @@ package com.provectus.kafka.ui.util.jsonschema; -import com.fasterxml.jackson.core.JsonProcessingException; import com.fasterxml.jackson.databind.ObjectMapper; -import com.github.fge.jsonschema.core.exceptions.ProcessingException; -import com.github.fge.jsonschema.core.report.ProcessingReport; -import com.github.fge.jsonschema.main.JsonSchemaFactory; -import io.confluent.kafka.schemaregistry.avro.AvroSchemaUtils; -import java.io.IOException; import java.net.URI; import java.net.URISyntaxException; +import lombok.SneakyThrows; import org.apache.avro.Schema; -import org.apache.avro.generic.GenericData; import org.junit.jupiter.api.Assertions; +import org.junit.jupiter.api.BeforeEach; import org.junit.jupiter.api.Test; -public class AvroJsonSchemaConverterTest { - @Test - public void avroConvertTest() throws URISyntaxException, JsonProcessingException { - final AvroJsonSchemaConverter converter = new AvroJsonSchemaConverter(); - URI basePath = new URI("http://example.com/"); +class AvroJsonSchemaConverterTest { + + private AvroJsonSchemaConverter converter; + private URI basePath; + + @BeforeEach + void init() throws URISyntaxException { + converter = new AvroJsonSchemaConverter(); + basePath = new URI("http://example.com/"); + } - Schema recordSchema = (new Schema.Parser()).parse( - " {" + @Test + void avroConvertTest() { + String avroSchema = + " {" + " \"type\": \"record\"," + " \"name\": \"Message\"," + " \"namespace\": \"com.provectus.kafka\"," @@ -76,45 +78,59 @@ public void avroConvertTest() throws URISyntaxException, JsonProcessingException + " }" + " }" + " ]" - + " }" - ); + + " }"; + String expectedJsonSchema = "{ " + + " \"$id\" : \"http://example.com/Message\", " + + " \"$schema\" : \"https://json-schema.org/draft/2020-12/schema\", " + + " \"type\" : \"object\", " + + " \"properties\" : { " + + " \"record\" : { \"$ref\" : \"#/definitions/com.provectus.kafka.InnerMessage\" } " + + " }, " + + " \"required\" : [ \"record\" ], " + + " \"definitions\" : { " + + " \"com.provectus.kafka.Message\" : { \"$ref\" : \"#\" }, " + + " \"com.provectus.kafka.InnerMessage\" : { " + + " \"type\" : \"object\", " + + " \"properties\" : { " + + " \"long_text\" : { " + + " \"oneOf\" : [ { " + + " \"type\" : \"null\" " + + " }, { " + + " \"type\" : \"object\", " + + " \"properties\" : { " + + " \"string\" : { " + + " \"type\" : \"string\" " + + " } " + + " } " + + " } ] " + + " }, " + + " \"array\" : { " + + " \"type\" : \"array\", " + + " \"items\" : { \"type\" : \"string\" } " + + " }, " + + " \"id\" : { \"type\" : \"integer\" }, " + + " \"text\" : { \"type\" : \"string\" }, " + + " \"map\" : { " + + " \"type\" : \"object\", " + + " \"additionalProperties\" : { \"type\" : \"integer\" } " + + " }, " + + " \"order\" : { " + + " \"enum\" : [ \"SPADES\", \"HEARTS\", \"DIAMONDS\", \"CLUBS\" ], " + + " \"type\" : \"string\" " + + " } " + + " }, " + + " \"required\" : [ \"id\", \"text\", \"order\", \"array\", \"map\" ] " + + " } " + + " } " + + "}"; - String expected = "{\"$id\":\"http://example.com/Message\"," - + "\"$schema\":\"https://json-schema.org/draft/2020-12/schema\"," - + "\"type\":\"object\",\"properties\":{\"record\":" - + "{\"$ref\":\"#/definitions/RecordInnerMessage\"}}," - + "\"required\":[\"record\"],\"definitions\":" - + "{\"RecordInnerMessage\":{\"type\":\"object\",\"" - + "properties\":{\"long_text\":{\"oneOf\":[{\"type\":\"null\"}," - + "{\"type\":\"object\",\"properties\":{\"string\":" - + "{\"type\":\"string\"}}}]},\"array\":{\"type\":\"array\",\"items\":" - + "{\"type\":\"string\"}},\"id\":{\"type\":\"integer\"},\"text\":" - + "{\"type\":\"string\"},\"map\":{\"type\":\"object\"," - + "\"additionalProperties\":{\"type\":\"integer\"}}," - + "\"order\":{\"enum\":[\"SPADES\",\"HEARTS\",\"DIAMONDS\",\"CLUBS\"]," - + "\"type\":\"string\"}}," - + "\"required\":[\"id\",\"text\",\"order\",\"array\",\"map\"]}}}"; - - final JsonSchema convertRecord = converter.convert(basePath, recordSchema); - - ObjectMapper om = new ObjectMapper(); - Assertions.assertEquals( - om.readTree(expected), - om.readTree( - convertRecord.toJson() - ) - ); - + convertAndCompare(expectedJsonSchema, avroSchema); } @Test - public void testNullableUnions() throws URISyntaxException, IOException, ProcessingException { - final AvroJsonSchemaConverter converter = new AvroJsonSchemaConverter(); - URI basePath = new URI("http://example.com/"); - final ObjectMapper objectMapper = new ObjectMapper(); - - Schema recordSchema = (new Schema.Parser()).parse( + void testNullableUnions() { + String avroSchema = " {" + " \"type\": \"record\"," + " \"name\": \"Message\"," @@ -138,38 +154,105 @@ public void testNullableUnions() throws URISyntaxException, IOException, Process + " \"default\": null" + " }" + " ]" - + " }" - ); - - final GenericData.Record record = new GenericData.Record(recordSchema); - record.put("text", "Hello world"); - record.put("value", 100L); - byte[] jsonBytes = AvroSchemaUtils.toJson(record); - String serialized = new String(jsonBytes); + + " }"; - String expected = + String expectedJsonSchema = "{\"$id\":\"http://example.com/Message\"," + "\"$schema\":\"https://json-schema.org/draft/2020-12/schema\"," + "\"type\":\"object\",\"properties\":{\"text\":" + "{\"oneOf\":[{\"type\":\"null\"},{\"type\":\"object\"," + "\"properties\":{\"string\":{\"type\":\"string\"}}}]},\"value\":" + "{\"oneOf\":[{\"type\":\"null\"},{\"type\":\"object\"," - + "\"properties\":{\"string\":{\"type\":\"string\"},\"long\":{\"type\":\"integer\"}}}]}}}"; + + "\"properties\":{\"string\":{\"type\":\"string\"},\"long\":{\"type\":\"integer\"}}}]}}," + + "\"definitions\" : { \"com.provectus.kafka.Message\" : { \"$ref\" : \"#\" }}}"; - final JsonSchema convert = converter.convert(basePath, recordSchema); - Assertions.assertEquals( - objectMapper.readTree(expected), - objectMapper.readTree(convert.toJson()) - ); + convertAndCompare(expectedJsonSchema, avroSchema); + } + @Test + void testRecordReferences() { + String avroSchema = + "{\n" + + " \"type\": \"record\", " + + " \"namespace\": \"n.s\", " + + " \"name\": \"RootMsg\", " + + " \"fields\":\n" + + " [ " + + " { " + + " \"name\": \"inner1\", " + + " \"type\": { " + + " \"type\": \"record\", " + + " \"name\": \"Inner\", " + + " \"fields\": [ { \"name\": \"f1\", \"type\": \"double\" } ] " + + " } " + + " }, " + + " { " + + " \"name\": \"inner2\", " + + " \"type\": { " + + " \"type\": \"record\", " + + " \"namespace\": \"n.s2\", " + + " \"name\": \"Inner\", " + + " \"fields\": " + + " [ { \"name\": \"f1\", \"type\": \"double\" } ] " + + " } " + + " }, " + + " { " + + " \"name\": \"refField\", " + + " \"type\": [ \"null\", \"Inner\", \"n.s2.Inner\", \"RootMsg\" ] " + + " } " + + " ] " + + "}"; + + String expectedJsonSchema = "{ " + + " \"$id\" : \"http://example.com/RootMsg\", " + + " \"$schema\" : \"https://json-schema.org/draft/2020-12/schema\", " + + " \"type\" : \"object\", " + + " \"properties\" : { " + + " \"inner1\" : { \"$ref\" : \"#/definitions/n.s.Inner\" }, " + + " \"inner2\" : { \"$ref\" : \"#/definitions/n.s2.Inner\" }, " + + " \"refField\" : { " + + " \"oneOf\" : [ " + + " { " + + " \"type\" : \"null\" " + + " }, " + + " { " + + " \"type\" : \"object\", " + + " \"properties\" : { " + + " \"n.s.RootMsg\" : { \"$ref\" : \"#/definitions/n.s.RootMsg\" }, " + + " \"n.s2.Inner\" : { \"$ref\" : \"#/definitions/n.s2.Inner\" }, " + + " \"n.s.Inner\" : { \"$ref\" : \"#/definitions/n.s.Inner\" } " + + " } " + + " } ] " + + " } " + + " }, " + + " \"required\" : [ \"inner1\", \"inner2\" ], " + + " \"definitions\" : { " + + " \"n.s.RootMsg\" : { \"$ref\" : \"#\" }, " + + " \"n.s2.Inner\" : { " + + " \"type\" : \"object\", " + + " \"properties\" : { \"f1\" : { \"type\" : \"number\" } }, " + + " \"required\" : [ \"f1\" ] " + + " }, " + + " \"n.s.Inner\" : { " + + " \"type\" : \"object\", " + + " \"properties\" : { \"f1\" : { \"type\" : \"number\" } }, " + + " \"required\" : [ \"f1\" ] " + + " } " + + " } " + + "}"; - final ProcessingReport validate = - JsonSchemaFactory.byDefault().getJsonSchema( - objectMapper.readTree(expected) - ).validate( - objectMapper.readTree(serialized) - ); + convertAndCompare(expectedJsonSchema, avroSchema); + } - Assertions.assertTrue(validate.isSuccess()); + @SneakyThrows + private void convertAndCompare(String expectedJsonSchema, String sourceAvroSchema) { + var parseAvroSchema = new Schema.Parser().parse(sourceAvroSchema); + var converted = converter.convert(basePath, parseAvroSchema).toJson(); + var objectMapper = new ObjectMapper(); + Assertions.assertEquals( + objectMapper.readTree(expectedJsonSchema), + objectMapper.readTree(converted) + ); } + } \ No newline at end of file
train
val
2022-09-27T17:52:59
"2022-02-08T09:40:55Z"
Haarolean
train
provectus/kafka-ui/2436_2656
provectus/kafka-ui
provectus/kafka-ui/2436
provectus/kafka-ui/2656
[ "timestamp(timedelta=1.0, similarity=0.9961329235646676)", "connected" ]
7db55d5acf29d1daab488ea959e40ff143054699
3f4791ff0a70418836a7de5f263674d128bd43a5
[]
[ "several apiHelpers?", "where's schemaForm?", "fixed", "added all forms ", "we don't need to remove navigation steps from test. please return mainPage.goTo()\n.goToSideMenu(CLUSTER_NAME, bla-bla", "we don't need to remove navigation steps from test. please return mainPage.goTo()\n.goToSideMenu(CLUSTER_NAME, bla-bla", "we don't need to remove navigation steps from test. please return mainPage.goTo()\n.goToSideMenu(CLUSTER_NAME, bla-bla", "we don't need to remove navigation steps from test. please return mainPage.goTo()\n.goToSideMenu(CLUSTER_NAME, bla-bla", "Done ", "Done ", "Done ", "Done " ]
"2022-09-27T13:21:34Z"
[ "type/enhancement", "scope/QA", "status/accepted" ]
[e2e] create screen facade class and refactor objects of page classes creation
- need to create this class with created screen class objects - inherit BaseTest from Facade - refactor objects of page classes creation
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/helpers/Helpers.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/Pages.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorCreateView.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorsView.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorCreateView.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorsView.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/SmokeTests.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/BaseTest.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/ConnectorsTests.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/SchemasTests.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/helpers/Helpers.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/helpers/Helpers.java deleted file mode 100644 index cda55a0cb25..00000000000 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/helpers/Helpers.java +++ /dev/null @@ -1,11 +0,0 @@ -package com.provectus.kafka.ui.helpers; - - - -public class Helpers { - public static final Helpers INSTANCE = new Helpers(); - - private Helpers(){} - - public ApiHelper apiHelper = new ApiHelper(); -} diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/Pages.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/Pages.java deleted file mode 100644 index 79a7652dd4d..00000000000 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/Pages.java +++ /dev/null @@ -1,53 +0,0 @@ -package com.provectus.kafka.ui.pages; - -import com.provectus.kafka.ui.pages.connector.ConnectorsList; -import com.provectus.kafka.ui.pages.connector.ConnectorsView; -import com.provectus.kafka.ui.pages.schema.SchemaRegistryList; -import com.provectus.kafka.ui.pages.topic.ProduceMessagePanel; -import com.provectus.kafka.ui.pages.topic.TopicView; -import com.provectus.kafka.ui.pages.topic.TopicsList; -import io.qameta.allure.Step; - -public class Pages { - - public static Pages INSTANCE = new Pages(); - - public MainPage mainPage = new MainPage(); - public TopicsList topicsList = new TopicsList(); - public TopicView topicView = new TopicView(); - public ProduceMessagePanel produceMessagePanel = new ProduceMessagePanel(); - public ConnectorsList connectorsList = new ConnectorsList(); - public ConnectorsView connectorsView = new ConnectorsView(); - public SchemaRegistryList schemaRegistry = new SchemaRegistryList(); - - @Step - public MainPage open() { - return openMainPage(); - } - - @Step - public MainPage openMainPage() { - return mainPage.goTo(); - } - - @Step - public TopicsList openTopicsList(String clusterName) { - return topicsList.goTo(clusterName); - } - - @Step - public TopicView openTopicView(String clusterName, String topicName) { - return topicView.goTo(clusterName, topicName); - } - - @Step - public ConnectorsList openConnectorsList(String clusterName) { - return connectorsList.goTo(clusterName); - } - - @Step - public ConnectorsView openConnectorsView(String clusterName, String connectorName) { - return connectorsView.goTo(clusterName, connectorName); - } - -} diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorCreateView.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorCreateView.java index 16faa760581..296dff5b1f0 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorCreateView.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorCreateView.java @@ -2,9 +2,7 @@ import com.codeborne.selenide.Condition; import com.codeborne.selenide.SelenideElement; -import com.provectus.kafka.ui.utilities.WaitUtils; import io.qameta.allure.Step; -import lombok.experimental.ExtensionMethod; import org.openqa.selenium.By; import static com.codeborne.selenide.Selenide.$; @@ -12,7 +10,6 @@ import static com.provectus.kafka.ui.utilities.WebUtils.javaExecutorClick; import static com.provectus.kafka.ui.utilities.screenshots.Screenshooter.log; -@ExtensionMethod(WaitUtils.class) public class ConnectorCreateView { SelenideElement nameField = $(By.xpath("//input[@name='name']")); diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorsView.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorsView.java index 48f05529e32..86c7f0dfdaa 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorsView.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorsView.java @@ -4,9 +4,7 @@ import com.codeborne.selenide.Selenide; import com.codeborne.selenide.SelenideElement; import com.provectus.kafka.ui.settings.Source; -import com.provectus.kafka.ui.utilities.WaitUtils; import io.qameta.allure.Step; -import lombok.experimental.ExtensionMethod; import org.openqa.selenium.By; import org.openqa.selenium.Keys; @@ -14,7 +12,6 @@ import static com.provectus.kafka.ui.utilities.WebUtils.javaExecutorClick; import static com.provectus.kafka.ui.utilities.screenshots.Screenshooter.log; -@ExtensionMethod(WaitUtils.class) public class ConnectorsView { private static final String path = "/ui/clusters/%s/connects/first/connectors/%s"; protected SelenideElement submitButton = $(By.xpath("//button[@type='submit']"));
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/SmokeTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/SmokeTests.java index ee2024d0e7d..c9c560c3edc 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/SmokeTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/SmokeTests.java @@ -1,8 +1,8 @@ package com.provectus.kafka.ui; import com.provectus.kafka.ui.base.BaseTest; -import com.provectus.kafka.ui.utilities.qaseIoUtils.enums.Status; import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.AutomationStatus; +import com.provectus.kafka.ui.utilities.qaseIoUtils.enums.Status; import io.qase.api.annotation.CaseId; import org.junit.jupiter.api.DisplayName; import org.junit.jupiter.api.Test; @@ -14,7 +14,7 @@ public class SmokeTests extends BaseTest { @CaseId(198) @DisplayName("main page should load") void mainPageLoads() { - pages.open() + mainPage.goTo() .waitUntilScreenReady(); compareScreenshots("main"); } diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/BaseTest.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/BaseTest.java index a69f3afc9dd..a864a36eb2c 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/BaseTest.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/BaseTest.java @@ -1,8 +1,6 @@ package com.provectus.kafka.ui.base; import com.codeborne.selenide.WebDriverRunner; -import com.provectus.kafka.ui.helpers.Helpers; -import com.provectus.kafka.ui.pages.Pages; import com.provectus.kafka.ui.utilities.qaseIoUtils.DisplayNameGenerator; import com.provectus.kafka.ui.utilities.qaseIoUtils.TestCaseGenerator; import com.provectus.kafka.ui.utilities.screenshots.Screenshooter; @@ -30,13 +28,11 @@ @Slf4j @DisplayNameGeneration(DisplayNameGenerator.class) -public class BaseTest { +public class BaseTest extends Facade { private static final String SELENIUM_IMAGE_NAME = "selenium/standalone-chrome:103.0"; private static final String SELENIARM_STANDALONE_CHROMIUM = "seleniarm/standalone-chromium:103.0"; protected static final String CLUSTER_NAME = "local"; - protected Pages pages = Pages.INSTANCE; - protected Helpers helpers = Helpers.INSTANCE; private final Screenshooter screenshooter = new Screenshooter(); diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java new file mode 100644 index 00000000000..8e9b1f149ab --- /dev/null +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java @@ -0,0 +1,31 @@ +package com.provectus.kafka.ui.base; + +import com.provectus.kafka.ui.helpers.ApiHelper; +import com.provectus.kafka.ui.pages.MainPage; +import com.provectus.kafka.ui.pages.connector.ConnectorCreateView; +import com.provectus.kafka.ui.pages.connector.ConnectorsList; +import com.provectus.kafka.ui.pages.connector.ConnectorsView; +import com.provectus.kafka.ui.pages.schema.SchemaCreateView; +import com.provectus.kafka.ui.pages.schema.SchemaEditView; +import com.provectus.kafka.ui.pages.schema.SchemaRegistryList; +import com.provectus.kafka.ui.pages.schema.SchemaView; +import com.provectus.kafka.ui.pages.topic.ProduceMessagePanel; +import com.provectus.kafka.ui.pages.topic.TopicCreateEditSettingsView; +import com.provectus.kafka.ui.pages.topic.TopicView; +import com.provectus.kafka.ui.pages.topic.TopicsList; + +public abstract class Facade { + protected MainPage mainPage = new MainPage(); + protected ApiHelper apiHelper = new ApiHelper(); + protected ConnectorCreateView connectorCreateView = new ConnectorCreateView(); + protected ConnectorsList connectorsList = new ConnectorsList(); + protected ConnectorsView connectorsView = new ConnectorsView(); + protected SchemaCreateView schemaCreateView = new SchemaCreateView(); + protected SchemaEditView schemaEditView = new SchemaEditView(); + protected SchemaView schemaView = new SchemaView(); + protected SchemaRegistryList schemaRegistryList = new SchemaRegistryList(); + protected ProduceMessagePanel produceMessagePanel = new ProduceMessagePanel(); + protected TopicCreateEditSettingsView topicCreateEditSettingsView = new TopicCreateEditSettingsView(); + protected TopicsList topicsList = new TopicsList(); + protected TopicView topicView = new TopicView(); +} diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/ConnectorsTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/ConnectorsTests.java index bc643fa5b3a..d99d08ffbd8 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/ConnectorsTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/ConnectorsTests.java @@ -1,12 +1,11 @@ package com.provectus.kafka.ui.tests; import com.provectus.kafka.ui.base.BaseTest; -import com.provectus.kafka.ui.helpers.Helpers; import com.provectus.kafka.ui.models.Connector; import com.provectus.kafka.ui.models.Topic; -import com.provectus.kafka.ui.utilities.qaseIoUtils.enums.Status; import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.AutomationStatus; import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.Suite; +import com.provectus.kafka.ui.utilities.qaseIoUtils.enums.Status; import io.qase.api.annotation.CaseId; import org.junit.jupiter.api.*; @@ -15,6 +14,7 @@ import static com.provectus.kafka.ui.utilities.FileUtils.getResourceAsString; +@TestInstance(TestInstance.Lifecycle.PER_CLASS) public class ConnectorsTests extends BaseTest { private static final long SUITE_ID = 10; private static final String SUITE_TITLE = "Kafka Connect"; @@ -40,14 +40,14 @@ public class ConnectorsTests extends BaseTest { .setConfig(getResourceAsString("config_for_create_connector_via_api.json")); @BeforeAll - public static void beforeAll() { + public void beforeAll() { TOPIC_LIST.addAll(List.of(TOPIC_FOR_CREATE, TOPIC_FOR_DELETE, TOPIC_FOR_UPDATE)); TOPIC_LIST.forEach(topic -> { - Helpers.INSTANCE.apiHelper.createTopic(CLUSTER_NAME, topic.getName()); - Helpers.INSTANCE.apiHelper.sendMessage(CLUSTER_NAME, topic); + apiHelper.createTopic(CLUSTER_NAME, topic.getName()); + apiHelper.sendMessage(CLUSTER_NAME, topic); }); CONNECTOR_LIST.addAll(List.of(CONNECTOR_FOR_DELETE, CONNECTOR_FOR_UPDATE)); - CONNECTOR_LIST.forEach(connector -> Helpers.INSTANCE.apiHelper + CONNECTOR_LIST.forEach(connector -> apiHelper .createConnector(CLUSTER_NAME, CONNECT_NAME, connector)); } @@ -60,14 +60,14 @@ public void createConnector() { Connector connectorForCreate = new Connector() .setName("sink_postgres_activities_e2e_checks") .setConfig(getResourceAsString("config_for_create_connector.json")); - pages.openConnectorsList(CLUSTER_NAME) + connectorsList.goTo(CLUSTER_NAME) .waitUntilScreenReady() .clickCreateConnectorButton() .waitUntilScreenReady() .setConnectorConfig(connectorForCreate.getName(), connectorForCreate.getConfig()); - pages.openConnectorsList(CLUSTER_NAME) + connectorsList.goTo(CLUSTER_NAME) .waitUntilScreenReady(); - Assertions.assertTrue(pages.connectorsList.isConnectorVisible(connectorForCreate.getName()),"isConnectorVisible()"); + Assertions.assertTrue(connectorsList.isConnectorVisible(connectorForCreate.getName()),"isConnectorVisible()"); CONNECTOR_LIST.add(connectorForCreate); } @@ -77,15 +77,14 @@ public void createConnector() { @CaseId(196) @Test public void updateConnector() { - pages.openConnectorsList(CLUSTER_NAME) + connectorsList.goTo(CLUSTER_NAME) .waitUntilScreenReady() .openConnector(CONNECTOR_FOR_UPDATE.getName()); - pages.connectorsView - .waitUntilScreenReady() + connectorsView.waitUntilScreenReady() .openConfigTab() .setConfig(CONNECTOR_FOR_UPDATE.getConfig()); - pages.openConnectorsList(CLUSTER_NAME); - Assertions.assertTrue(pages.connectorsList.isConnectorVisible(CONNECTOR_FOR_UPDATE.getName()),"isConnectorVisible()"); + connectorsList.goTo(CLUSTER_NAME); + Assertions.assertTrue(connectorsList.isConnectorVisible(CONNECTOR_FOR_UPDATE.getName()),"isConnectorVisible()"); } @DisplayName("should delete connector") @@ -94,19 +93,19 @@ public void updateConnector() { @CaseId(195) @Test public void deleteConnector() { - pages.openConnectorsList(CLUSTER_NAME) + connectorsList.goTo(CLUSTER_NAME) .waitUntilScreenReady() .openConnector(CONNECTOR_FOR_DELETE.getName()); - pages.connectorsView.clickDeleteButton(); - pages.openConnectorsList(CLUSTER_NAME); - Assertions.assertFalse(pages.connectorsList.isConnectorVisible(CONNECTOR_FOR_DELETE.getName()),"isConnectorVisible()"); + connectorsView.clickDeleteButton(); + connectorsList.goTo(CLUSTER_NAME); + Assertions.assertFalse(connectorsList.isConnectorVisible(CONNECTOR_FOR_DELETE.getName()),"isConnectorVisible()"); CONNECTOR_LIST.remove(CONNECTOR_FOR_DELETE); } @AfterAll - public static void afterAll() { + public void afterAll() { CONNECTOR_LIST.forEach(connector -> - Helpers.INSTANCE.apiHelper.deleteConnector(CLUSTER_NAME, CONNECT_NAME, connector.getName())); - TOPIC_LIST.forEach(topic -> Helpers.INSTANCE.apiHelper.deleteTopic(CLUSTER_NAME, topic.getName())); + apiHelper.deleteConnector(CLUSTER_NAME, CONNECT_NAME, connector.getName())); + TOPIC_LIST.forEach(topic -> apiHelper.deleteTopic(CLUSTER_NAME, topic.getName())); } } diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/SchemasTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/SchemasTests.java index 02588bcc100..563cc343cac 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/SchemasTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/SchemasTests.java @@ -2,14 +2,12 @@ import com.provectus.kafka.ui.api.model.CompatibilityLevel; import com.provectus.kafka.ui.base.BaseTest; -import com.provectus.kafka.ui.helpers.Helpers; import com.provectus.kafka.ui.models.Schema; import com.provectus.kafka.ui.pages.MainPage; -import com.provectus.kafka.ui.pages.schema.SchemaEditView; import com.provectus.kafka.ui.pages.schema.SchemaView; -import com.provectus.kafka.ui.utilities.qaseIoUtils.enums.Status; import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.AutomationStatus; import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.Suite; +import com.provectus.kafka.ui.utilities.qaseIoUtils.enums.Status; import io.qase.api.annotation.CaseId; import lombok.SneakyThrows; import org.junit.jupiter.api.*; @@ -19,7 +17,7 @@ import static com.provectus.kafka.ui.utilities.FileUtils.fileToString; - +@TestInstance(TestInstance.Lifecycle.PER_CLASS) @TestMethodOrder(MethodOrderer.OrderAnnotation.class) public class SchemasTests extends BaseTest { private static final long SUITE_ID = 11; @@ -31,9 +29,9 @@ public class SchemasTests extends BaseTest { @BeforeAll @SneakyThrows - public static void beforeAll() { + public void beforeAll() { SCHEMA_LIST.addAll(List.of(AVRO_API, JSON_API, PROTOBUF_API)); - SCHEMA_LIST.forEach(schema -> Helpers.INSTANCE.apiHelper.createSchema(CLUSTER_NAME, schema)); + SCHEMA_LIST.forEach(schema -> apiHelper.createSchema(CLUSTER_NAME, schema)); } @DisplayName("should create AVRO schema") @@ -44,17 +42,17 @@ public static void beforeAll() { @Order(1) void createSchemaAvro() { Schema schemaAvro = Schema.createSchemaAvro(); - pages.openMainPage() + mainPage.goTo() .goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.SCHEMA_REGISTRY); - pages.schemaRegistry.clickCreateSchema() + schemaRegistryList.clickCreateSchema() .setSubjectName(schemaAvro.getName()) .setSchemaField(fileToString(schemaAvro.getValuePath())) .selectSchemaTypeFromDropdown(schemaAvro.getType()) .clickSubmit() .waitUntilScreenReady(); - pages.mainPage + mainPage.goTo() .goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.SCHEMA_REGISTRY); - Assertions.assertTrue(pages.schemaRegistry.isSchemaVisible(schemaAvro.getName()),"isSchemaVisible()"); + Assertions.assertTrue(schemaRegistryList.isSchemaVisible(schemaAvro.getName()),"isSchemaVisible()"); SCHEMA_LIST.add(schemaAvro); } @@ -66,13 +64,13 @@ void createSchemaAvro() { @Order(2) void updateSchemaAvro() { AVRO_API.setValuePath(System.getProperty("user.dir") + "/src/main/resources/testData/schema_avro_for_update.json"); - pages.openMainPage() + mainPage.goTo() .goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.SCHEMA_REGISTRY); - pages.schemaRegistry.openSchema(AVRO_API.getName()) + schemaRegistryList.openSchema(AVRO_API.getName()) .waitUntilScreenReady() .openEditSchema(); - Assertions.assertTrue(new SchemaEditView().isSchemaDropDownDisabled(),"isSchemaDropDownDisabled()"); - new SchemaEditView().selectCompatibilityLevelFromDropdown(CompatibilityLevel.CompatibilityEnum.NONE) + Assertions.assertTrue(schemaEditView.isSchemaDropDownDisabled(),"isSchemaDropDownDisabled()"); + schemaEditView.selectCompatibilityLevelFromDropdown(CompatibilityLevel.CompatibilityEnum.NONE) .setNewSchemaValue(fileToString(AVRO_API.getValuePath())) .clickSubmit() .waitUntilScreenReady(); @@ -86,12 +84,12 @@ void updateSchemaAvro() { @Test @Order(3) void deleteSchemaAvro() { - pages.openMainPage() + mainPage.goTo() .goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.SCHEMA_REGISTRY); - pages.schemaRegistry.openSchema(AVRO_API.getName()) + schemaRegistryList.openSchema(AVRO_API.getName()) .waitUntilScreenReady() .removeSchema(); - Assertions.assertFalse(pages.schemaRegistry.isSchemaVisible(AVRO_API.getName()),"isSchemaVisible()"); + Assertions.assertFalse(schemaRegistryList.isSchemaVisible(AVRO_API.getName()),"isSchemaVisible()"); SCHEMA_LIST.remove(AVRO_API); } @@ -103,17 +101,16 @@ void deleteSchemaAvro() { @Order(4) void createSchemaJson() { Schema schemaJson = Schema.createSchemaJson(); - pages.openMainPage() + mainPage.goTo() .goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.SCHEMA_REGISTRY); - pages.schemaRegistry.clickCreateSchema() + schemaRegistryList.clickCreateSchema() .setSubjectName(schemaJson.getName()) .setSchemaField(fileToString(schemaJson.getValuePath())) .selectSchemaTypeFromDropdown(schemaJson.getType()) .clickSubmit() .waitUntilScreenReady(); - pages.mainPage - .goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.SCHEMA_REGISTRY); - Assertions.assertTrue(pages.schemaRegistry.isSchemaVisible(schemaJson.getName()),"isSchemaVisible()"); + mainPage.goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.SCHEMA_REGISTRY); + Assertions.assertTrue(schemaRegistryList.isSchemaVisible(schemaJson.getName()),"isSchemaVisible()"); SCHEMA_LIST.add(schemaJson); } @@ -124,12 +121,12 @@ void createSchemaJson() { @Test @Order(5) void deleteSchemaJson() { - pages.openMainPage() + mainPage.goTo() .goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.SCHEMA_REGISTRY); - pages.schemaRegistry.openSchema(JSON_API.getName()) + schemaRegistryList.openSchema(JSON_API.getName()) .waitUntilScreenReady() .removeSchema(); - Assertions.assertFalse(pages.schemaRegistry.isSchemaVisible(JSON_API.getName()),"isSchemaVisible()"); + Assertions.assertFalse(schemaRegistryList.isSchemaVisible(JSON_API.getName()),"isSchemaVisible()"); SCHEMA_LIST.remove(JSON_API); } @@ -141,17 +138,16 @@ void deleteSchemaJson() { @Order(6) void createSchemaProtobuf() { Schema schemaProtobuf = Schema.createSchemaProtobuf(); - pages.openMainPage() + mainPage.goTo() .goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.SCHEMA_REGISTRY); - pages.schemaRegistry.clickCreateSchema() + schemaRegistryList.clickCreateSchema() .setSubjectName(schemaProtobuf.getName()) .setSchemaField(fileToString(schemaProtobuf.getValuePath())) .selectSchemaTypeFromDropdown(schemaProtobuf.getType()) .clickSubmit() .waitUntilScreenReady(); - pages.mainPage - .goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.SCHEMA_REGISTRY); - Assertions.assertTrue(pages.schemaRegistry.isSchemaVisible(schemaProtobuf.getName()),"isSchemaVisible()"); + mainPage.goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.SCHEMA_REGISTRY); + Assertions.assertTrue(schemaRegistryList.isSchemaVisible(schemaProtobuf.getName()),"isSchemaVisible()"); SCHEMA_LIST.add(schemaProtobuf); } @@ -162,17 +158,17 @@ void createSchemaProtobuf() { @Test @Order(7) void deleteSchemaProtobuf() { - pages.openMainPage() + mainPage.goTo() .goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.SCHEMA_REGISTRY); - pages.schemaRegistry.openSchema(PROTOBUF_API.getName()) + schemaRegistryList.openSchema(PROTOBUF_API.getName()) .waitUntilScreenReady() .removeSchema(); - Assertions.assertFalse(pages.schemaRegistry.isSchemaVisible(PROTOBUF_API.getName()),"isSchemaVisible()"); + Assertions.assertFalse(schemaRegistryList.isSchemaVisible(PROTOBUF_API.getName()),"isSchemaVisible()"); SCHEMA_LIST.remove(PROTOBUF_API); } @AfterAll - public static void afterAll() { - SCHEMA_LIST.forEach(schema -> Helpers.INSTANCE.apiHelper.deleteSchema(CLUSTER_NAME, schema.getName())); + public void afterAll() { + SCHEMA_LIST.forEach(schema -> apiHelper.deleteSchema(CLUSTER_NAME, schema.getName())); } } diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java index 5c6b0b86c18..bc4f6d4b658 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java @@ -1,14 +1,12 @@ package com.provectus.kafka.ui.tests; import com.provectus.kafka.ui.base.BaseTest; -import com.provectus.kafka.ui.helpers.Helpers; import com.provectus.kafka.ui.models.Topic; import com.provectus.kafka.ui.pages.MainPage; -import com.provectus.kafka.ui.pages.topic.TopicCreateEditSettingsView; import com.provectus.kafka.ui.pages.topic.TopicView; -import com.provectus.kafka.ui.utilities.qaseIoUtils.enums.Status; import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.AutomationStatus; import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.Suite; +import com.provectus.kafka.ui.utilities.qaseIoUtils.enums.Status; import io.qase.api.annotation.CaseId; import org.assertj.core.api.SoftAssertions; import org.junit.jupiter.api.*; @@ -18,6 +16,7 @@ import static com.provectus.kafka.ui.utilities.FileUtils.fileToString; +@TestInstance(TestInstance.Lifecycle.PER_CLASS) public class TopicTests extends BaseTest { private static final long SUITE_ID = 2; private static final String SUITE_TITLE = "Topics"; @@ -33,9 +32,9 @@ public class TopicTests extends BaseTest { private static final List<Topic> TOPIC_LIST = new ArrayList<>(); @BeforeAll - public static void beforeAll() { + public void beforeAll() { TOPIC_LIST.addAll(List.of(TOPIC_FOR_UPDATE, TOPIC_FOR_DELETE)); - TOPIC_LIST.forEach(topic -> Helpers.INSTANCE.apiHelper.createTopic(CLUSTER_NAME, topic.getName())); + TOPIC_LIST.forEach(topic -> apiHelper.createTopic(CLUSTER_NAME, topic.getName())); } @DisplayName("should create a topic") @@ -45,15 +44,15 @@ public static void beforeAll() { @Test public void createTopic() { Topic topicToCreate = new Topic().setName("new-topic"); - pages.open() + mainPage.goTo() .goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.TOPICS); - pages.topicsList.pressCreateNewTopic() + topicsList.pressCreateNewTopic() .setTopicName(topicToCreate.getName()) .sendData() .waitUntilScreenReady(); - pages.open() + mainPage.goTo() .goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.TOPICS); - Assertions.assertTrue(pages.topicsList.isTopicVisible(topicToCreate.getName()),"isTopicVisible"); + Assertions.assertTrue(topicsList.isTopicVisible(topicToCreate.getName()),"isTopicVisible"); TOPIC_LIST.add(topicToCreate); } @@ -64,9 +63,9 @@ public void createTopic() { @CaseId(197) @Test public void updateTopic() { - pages.openTopicsList(CLUSTER_NAME) + topicsList.goTo(CLUSTER_NAME) .waitUntilScreenReady(); - pages.topicsList.openTopic(TOPIC_FOR_UPDATE.getName()) + topicsList.openTopic(TOPIC_FOR_UPDATE.getName()) .waitUntilScreenReady() .openEditSettings() .selectCleanupPolicy(TOPIC_FOR_UPDATE.getCompactPolicyValue()) @@ -76,16 +75,16 @@ public void updateTopic() { .setMaxMessageBytes(TOPIC_FOR_UPDATE.getMaxMessageBytes()) .sendData() .waitUntilScreenReady(); - pages.openTopicsList(CLUSTER_NAME) + topicsList.goTo(CLUSTER_NAME) .waitUntilScreenReady(); - pages.topicsList.openTopic(TOPIC_FOR_UPDATE.getName()) + topicsList.openTopic(TOPIC_FOR_UPDATE.getName()) .waitUntilScreenReady() .openEditSettings(); SoftAssertions softly = new SoftAssertions(); - softly.assertThat(new TopicCreateEditSettingsView().getCleanupPolicy()).as("Cleanup Policy").isEqualTo(TOPIC_FOR_UPDATE.getCompactPolicyValue()); - softly.assertThat(new TopicCreateEditSettingsView().getTimeToRetain()).as("Time to retain").isEqualTo(TOPIC_FOR_UPDATE.getTimeToRetainData()); - softly.assertThat(new TopicCreateEditSettingsView().getMaxSizeOnDisk()).as("Max size on disk").isEqualTo(TOPIC_FOR_UPDATE.getMaxSizeOnDisk()); - softly.assertThat(new TopicCreateEditSettingsView().getMaxMessageBytes()).as("Max message bytes").isEqualTo(TOPIC_FOR_UPDATE.getMaxMessageBytes()); + softly.assertThat(topicCreateEditSettingsView.getCleanupPolicy()).as("Cleanup Policy").isEqualTo(TOPIC_FOR_UPDATE.getCompactPolicyValue()); + softly.assertThat(topicCreateEditSettingsView.getTimeToRetain()).as("Time to retain").isEqualTo(TOPIC_FOR_UPDATE.getTimeToRetainData()); + softly.assertThat(topicCreateEditSettingsView.getMaxSizeOnDisk()).as("Max size on disk").isEqualTo(TOPIC_FOR_UPDATE.getMaxSizeOnDisk()); + softly.assertThat(topicCreateEditSettingsView.getMaxMessageBytes()).as("Max message bytes").isEqualTo(TOPIC_FOR_UPDATE.getMaxMessageBytes()); softly.assertAll(); } @@ -95,14 +94,14 @@ public void updateTopic() { @CaseId(207) @Test public void deleteTopic() { - pages.openTopicsList(CLUSTER_NAME) + topicsList.goTo(CLUSTER_NAME) .waitUntilScreenReady() .openTopic(TOPIC_FOR_DELETE.getName()) .waitUntilScreenReady() .deleteTopic(); - pages.openTopicsList(CLUSTER_NAME) + topicsList.goTo(CLUSTER_NAME) .waitUntilScreenReady(); - Assertions.assertFalse(pages.topicsList.isTopicVisible(TOPIC_FOR_DELETE.getName()),"isTopicVisible"); + Assertions.assertFalse(topicsList.isTopicVisible(TOPIC_FOR_DELETE.getName()),"isTopicVisible"); TOPIC_LIST.remove(TOPIC_FOR_DELETE); } @@ -112,7 +111,7 @@ public void deleteTopic() { @CaseId(222) @Test void produceMessage() { - pages.openTopicsList(CLUSTER_NAME) + topicsList.goTo(CLUSTER_NAME) .waitUntilScreenReady() .openTopic(TOPIC_FOR_UPDATE.getName()) .waitUntilScreenReady() @@ -122,13 +121,13 @@ void produceMessage() { .setKeyField(TOPIC_FOR_UPDATE.getMessageKey()) .submitProduceMessage(); SoftAssertions softly = new SoftAssertions(); - softly.assertThat(pages.topicView.isKeyMessageVisible((TOPIC_FOR_UPDATE.getMessageKey()))).withFailMessage("isKeyMessageVisible()").isTrue(); - softly.assertThat(pages.topicView.isContentMessageVisible((TOPIC_FOR_UPDATE.getMessageContent()).trim())).withFailMessage("isContentMessageVisible()").isTrue(); + softly.assertThat(topicView.isKeyMessageVisible((TOPIC_FOR_UPDATE.getMessageKey()))).withFailMessage("isKeyMessageVisible()").isTrue(); + softly.assertThat(topicView.isContentMessageVisible((TOPIC_FOR_UPDATE.getMessageContent()).trim())).withFailMessage("isContentMessageVisible()").isTrue(); softly.assertAll(); } @AfterAll - public static void afterAll() { - TOPIC_LIST.forEach(topic -> Helpers.INSTANCE.apiHelper.deleteTopic(CLUSTER_NAME, topic.getName())); + public void afterAll() { + TOPIC_LIST.forEach(topic -> apiHelper.deleteTopic(CLUSTER_NAME, topic.getName())); } }
val
val
2022-09-28T13:56:03
"2022-08-12T06:50:01Z"
VladSenyuta
train
provectus/kafka-ui/2536_2662
provectus/kafka-ui
provectus/kafka-ui/2536
provectus/kafka-ui/2662
[ "timestamp(timedelta=1.0, similarity=0.9601228571391652)", "connected" ]
3f4791ff0a70418836a7de5f263674d128bd43a5
c9613448242ff85c5f83361ec0c381ef39c37e58
[]
[ "pls use BytesFormatted here", "done" ]
"2022-09-27T16:57:48Z"
[ "good first issue", "scope/frontend", "status/accepted", "type/chore", "hacktoberfest" ]
Add "Total size" of Partition within Topic's statistics
**Describe the bug** Add "Total size" of Partition within Topic's statistics **Set up** https://www.kafka-ui.provectus.io/ **Steps to Reproduce** <!-- We'd like you to provide an example setup (via docker-compose, helm, etc.) to reproduce the problem, especially with a complex setups. --> Steps to reproduce the behavior: 1. Navigate to Topics 2. Open the Topic 3. Turn to Statistics tab 4. Open the Partition **Expected behavior** Total size should be displayed **Screenshots** <img width="1717" alt="total size" src="https://user-images.githubusercontent.com/104780608/188836260-3d02590b-23b2-4b89-b440-4d0c29e3e87e.png">
[ "kafka-ui-react-app/src/components/Topics/Topic/Statistics/PartitionInfoRow.tsx" ]
[ "kafka-ui-react-app/src/components/Topics/Topic/Statistics/PartitionInfoRow.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Statistics/PartitionInfoRow.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Statistics/PartitionInfoRow.tsx index 546f3e619d6..698019f2c44 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Statistics/PartitionInfoRow.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Statistics/PartitionInfoRow.tsx @@ -27,7 +27,6 @@ const PartitionInfoRow: React.FC<{ row: Row<TopicAnalysisStats> }> = ({ keySize, valueSize, } = row.original; - return ( <S.PartitionInfo> <div> @@ -35,6 +34,8 @@ const PartitionInfoRow: React.FC<{ row: Row<TopicAnalysisStats> }> = ({ <List> <Label>Total message</Label> <span>{totalMsgs}</span> + <Label>Total size</Label> + <BytesFormatted value={(keySize?.sum || 0) + (valueSize?.sum || 0)} /> <Label>Min. timestamp</Label> <span>{formatTimestamp(minTimestamp)}</span> <Label>Max. timestamp</Label>
null
test
val
2022-09-29T13:34:48
"2022-09-07T08:56:31Z"
armenuikafka
train
provectus/kafka-ui/2534_2663
provectus/kafka-ui
provectus/kafka-ui/2534
provectus/kafka-ui/2663
[ "connected", "timestamp(timedelta=1.0, similarity=0.9738042865951876)" ]
2f786c080b78b39740a68297879775c4fc31e295
a1d14ab4304a4b5d51a8537ee38d128aeaaa2719
[]
[ "pls use existing palete", "done", "```suggestion\r\n Change these parameters only if you are absolutely sure what you are\r\n```" ]
"2022-09-27T20:11:32Z"
[ "good first issue", "scope/frontend", "status/accepted", "type/chore", "hacktoberfest" ]
Add warning message for Danger zone in case of editing the Topic
Warning message should be displayed for Danger zone with editing the Topic **Describe the bug** Warning message should be displayed for Danger zone with editing the Topic **Set up** https://www.kafka-ui.provectus.io/ **Steps to Reproduce** 1. Navigate to Topics 2. Open any Topic 3. Edit settings 4. Check Danger zone **Expected behavior** Should be displayed warning message:"Change these parameters only if you are absolutely sure that you are doing." **Screenshots** <img width="1717" alt="warning message edit" src="https://user-images.githubusercontent.com/104780608/188640196-220b54ee-1c43-47fc-bc82-a7532f584e3d.png">
[ "kafka-ui-react-app/src/components/Topics/Topic/Edit/DangerZone/DangerZone.styled.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Edit/DangerZone/DangerZone.tsx", "kafka-ui-react-app/src/theme/theme.ts" ]
[ "kafka-ui-react-app/src/components/Topics/Topic/Edit/DangerZone/DangerZone.styled.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Edit/DangerZone/DangerZone.tsx", "kafka-ui-react-app/src/theme/theme.ts" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Edit/DangerZone/DangerZone.styled.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Edit/DangerZone/DangerZone.styled.tsx index b9ad51a669e..b5741c67e19 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Edit/DangerZone/DangerZone.styled.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Edit/DangerZone/DangerZone.styled.tsx @@ -15,11 +15,15 @@ export const Wrapper = styled.div` `; export const Title = styled.h1` - color: ${({ theme }) => theme.dangerZone.color}; + color: ${({ theme }) => theme.dangerZone.color.title}; font-size: 20px; padding-bottom: 16px; `; - +export const Warning = styled.div` + color: ${({ theme }) => theme.dangerZone.color.warningMessage}; + font-size: 12px; + padding-bottom: 16px; +`; export const Form = styled.form` display: flex; align-items: flex-end; diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Edit/DangerZone/DangerZone.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Edit/DangerZone/DangerZone.tsx index 49f1d3574a4..59016d1d7a4 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Edit/DangerZone/DangerZone.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Edit/DangerZone/DangerZone.tsx @@ -81,6 +81,10 @@ const DangerZone: React.FC<DangerZoneProps> = ({ return ( <S.Wrapper> <S.Title>Danger Zone</S.Title> + <S.Warning> + Change these parameters only if you are absolutely sure what you are + doing. + </S.Warning> <div> <FormProvider {...partitionsMethods}> <S.Form diff --git a/kafka-ui-react-app/src/theme/theme.ts b/kafka-ui-react-app/src/theme/theme.ts index d05ad54cb94..19c082c2840 100644 --- a/kafka-ui-react-app/src/theme/theme.ts +++ b/kafka-ui-react-app/src/theme/theme.ts @@ -518,7 +518,10 @@ const theme = { }, dangerZone: { borderColor: Colors.neutral[10], - color: Colors.red[50], + color: { + title: Colors.red[50], + warningMessage: Colors.neutral[50], + }, }, configList: { color: Colors.neutral[30],
null
test
val
2022-09-30T11:21:54
"2022-09-06T12:53:50Z"
armenuikafka
train
provectus/kafka-ui/2677_2687
provectus/kafka-ui
provectus/kafka-ui/2677
provectus/kafka-ui/2687
[ "keyword_pr_to_issue", "timestamp(timedelta=1.0, similarity=0.9557199577774567)" ]
a1d14ab4304a4b5d51a8537ee38d128aeaaa2719
55c8d6ac2d4bdfb9721c79d4894732c86e0380d9
[]
[]
"2022-09-30T19:54:16Z"
[ "good first issue", "scope/frontend", "status/accepted", "type/chore", "hacktoberfest" ]
Topic list: Rename "total partitions" to "partitions"
<img width="299" alt="image" src="https://user-images.githubusercontent.com/1494347/193238757-d6f3d11f-05dd-4d77-bd02-972ed95cfb3a.png">
[ "kafka-ui-react-app/src/components/Topics/List/TopicTable.tsx" ]
[ "kafka-ui-react-app/src/components/Topics/List/TopicTable.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/List/TopicTable.tsx b/kafka-ui-react-app/src/components/Topics/List/TopicTable.tsx index a855435c876..dcafa61b915 100644 --- a/kafka-ui-react-app/src/components/Topics/List/TopicTable.tsx +++ b/kafka-ui-react-app/src/components/Topics/List/TopicTable.tsx @@ -42,7 +42,7 @@ const TopicTable: React.FC = () => { }, { id: TopicColumnsToSort.TOTAL_PARTITIONS, - header: 'Total Partitions', + header: 'Partitions', accessorKey: 'partitionCount', }, {
null
test
val
2022-09-30T15:04:40
"2022-09-30T09:23:10Z"
Haarolean
train
provectus/kafka-ui/2076_2689
provectus/kafka-ui
provectus/kafka-ui/2076
provectus/kafka-ui/2689
[ "keyword_pr_to_issue" ]
80eb2dccfea9b15b64e927493fbbc5158549c935
7d5b7de992da9ddb1a018a6a61db66c4dcb204aa
[ "Implement sorting by: name, connect, type and status.", "Sorting by Name, Connect, Type, Plugin, Status implemented." ]
[ "Can we use existing `NewTable/TagCell` component?", "```suggestion\r\nconst TopicsCell: React.FC<CellContext<FullConnectorInfo, unknown>> = ({\r\n```", "useMemo is redundant here", "```suggestion\r\n if (!tasksCount) {\r\n return null;\r\n }\r\n \r\n return (tasksCount ? \r\n <>\r\n {tasksCount - (failedTasksCount || 0)} of {tasksCount}\r\n </>\r\n );\r\n```", "Pls use styled components attr to change link color. Or update Tag styles to show link correctly" ]
"2022-10-01T02:20:41Z"
[ "type/enhancement", "good first issue", "scope/backend", "scope/frontend", "status/accepted", "hacktoberfest", "status/pending-frontend" ]
Implement connectors sorting
1. Migrate table to new table component 2. Implement sorting 3. Related issue https://github.com/provectus/kafka-ui/issues/2325 The list is unsorted at all <img width="299" alt="image" src="https://user-images.githubusercontent.com/1494347/171179908-5447271c-e17d-4921-abfe-3a7e4239756a.png">
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java", "kafka-ui-react-app/src/components/Connect/List/List.styled.ts", "kafka-ui-react-app/src/components/Connect/List/List.tsx", "kafka-ui-react-app/src/components/Connect/List/ListItem.tsx", "kafka-ui-react-app/src/components/Connect/List/__tests__/List.spec.tsx", "kafka-ui-react-app/src/components/Connect/List/__tests__/ListItem.spec.tsx" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java", "kafka-ui-react-app/src/components/Connect/List/ActionsCell.tsx", "kafka-ui-react-app/src/components/Connect/List/List.styled.ts", "kafka-ui-react-app/src/components/Connect/List/List.tsx", "kafka-ui-react-app/src/components/Connect/List/RunningTasksCell.tsx", "kafka-ui-react-app/src/components/Connect/List/TopicsCell.tsx", "kafka-ui-react-app/src/components/Connect/List/__tests__/List.spec.tsx" ]
[]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java index b5986c7919e..89a73ff6931 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java @@ -28,14 +28,14 @@ public ConnectorCreateForm clickCreateConnectorButton() { @Step public KafkaConnectList openConnector(String connectorName) { - $(By.linkText(connectorName)).click(); + $x("//tbody//td[1][text()='" + connectorName + "']").shouldBe(Condition.enabled).click(); return this; } @Step public boolean isConnectorVisible(String connectorName) { $(By.xpath("//table")).shouldBe(Condition.visible); - return isVisible($x("//tbody//td[1]//a[text()='" + connectorName + "']")); + return isVisible($x("//tbody//td[1][text()='" + connectorName + "']")); } @Step diff --git a/kafka-ui-react-app/src/components/Connect/List/ActionsCell.tsx b/kafka-ui-react-app/src/components/Connect/List/ActionsCell.tsx new file mode 100644 index 00000000000..30b3df8a56d --- /dev/null +++ b/kafka-ui-react-app/src/components/Connect/List/ActionsCell.tsx @@ -0,0 +1,43 @@ +import React from 'react'; +import { FullConnectorInfo } from 'generated-sources'; +import { CellContext } from '@tanstack/react-table'; +import { ClusterNameRoute } from 'lib/paths'; +import useAppParams from 'lib/hooks/useAppParams'; +import { Dropdown, DropdownItem } from 'components/common/Dropdown'; +import { useDeleteConnector } from 'lib/hooks/api/kafkaConnect'; +import { useConfirm } from 'lib/hooks/useConfirm'; + +const ActionsCell: React.FC<CellContext<FullConnectorInfo, unknown>> = ({ + row, +}) => { + const { connect, name } = row.original; + + const { clusterName } = useAppParams<ClusterNameRoute>(); + + const confirm = useConfirm(); + const deleteMutation = useDeleteConnector({ + clusterName, + connectName: connect, + connectorName: name, + }); + + const handleDelete = () => { + confirm( + <> + Are you sure want to remove <b>{name}</b> connector? + </>, + async () => { + await deleteMutation.mutateAsync(); + } + ); + }; + return ( + <Dropdown> + <DropdownItem onClick={handleDelete} danger> + Remove Connector + </DropdownItem> + </Dropdown> + ); +}; + +export default ActionsCell; diff --git a/kafka-ui-react-app/src/components/Connect/List/List.styled.ts b/kafka-ui-react-app/src/components/Connect/List/List.styled.ts index f0e2631d2c1..799915fcb11 100644 --- a/kafka-ui-react-app/src/components/Connect/List/List.styled.ts +++ b/kafka-ui-react-app/src/components/Connect/List/List.styled.ts @@ -3,4 +3,10 @@ import styled from 'styled-components'; export const TagsWrapper = styled.div` display: flex; flex-wrap: wrap; + span { + color: rgb(76, 76, 255) !important; + &:hover { + color: rgb(23, 23, 207) !important; + } + } `; diff --git a/kafka-ui-react-app/src/components/Connect/List/List.tsx b/kafka-ui-react-app/src/components/Connect/List/List.tsx index 75a782fa0a6..b5935e7bab2 100644 --- a/kafka-ui-react-app/src/components/Connect/List/List.tsx +++ b/kafka-ui-react-app/src/components/Connect/List/List.tsx @@ -1,14 +1,18 @@ import React from 'react'; import useAppParams from 'lib/hooks/useAppParams'; -import { ClusterNameRoute } from 'lib/paths'; -import { Table } from 'components/common/table/Table/Table.styled'; -import TableHeaderCell from 'components/common/table/TableHeaderCell/TableHeaderCell'; +import { clusterConnectConnectorPath, ClusterNameRoute } from 'lib/paths'; +import Table, { TagCell } from 'components/common/NewTable'; +import { FullConnectorInfo } from 'generated-sources'; import { useConnectors } from 'lib/hooks/api/kafkaConnect'; -import { useSearchParams } from 'react-router-dom'; +import { ColumnDef } from '@tanstack/react-table'; +import { useNavigate, useSearchParams } from 'react-router-dom'; -import ListItem from './ListItem'; +import ActionsCell from './ActionsCell'; +import TopicsCell from './TopicsCell'; +import RunningTasksCell from './RunningTasksCell'; const List: React.FC = () => { + const navigate = useNavigate(); const { clusterName } = useAppParams<ClusterNameRoute>(); const [searchParams] = useSearchParams(); const { data: connectors } = useConnectors( @@ -16,35 +20,30 @@ const List: React.FC = () => { searchParams.get('q') || '' ); + const columns = React.useMemo<ColumnDef<FullConnectorInfo>[]>( + () => [ + { header: 'Name', accessorKey: 'name' }, + { header: 'Connect', accessorKey: 'connect' }, + { header: 'Type', accessorKey: 'type' }, + { header: 'Plugin', accessorKey: 'connectorClass' }, + { header: 'Topics', cell: TopicsCell }, + { header: 'Status', accessorKey: 'status.state', cell: TagCell }, + { header: 'Running Tasks', cell: RunningTasksCell }, + { header: '', id: 'action', cell: ActionsCell }, + ], + [] + ); + return ( - <Table isFullwidth> - <thead> - <tr> - <TableHeaderCell title="Name" /> - <TableHeaderCell title="Connect" /> - <TableHeaderCell title="Type" /> - <TableHeaderCell title="Plugin" /> - <TableHeaderCell title="Topics" /> - <TableHeaderCell title="Status" /> - <TableHeaderCell title="Running Tasks" /> - <TableHeaderCell> </TableHeaderCell> - </tr> - </thead> - <tbody> - {(!connectors || connectors.length) === 0 && ( - <tr> - <td colSpan={10}>No connectors found</td> - </tr> - )} - {connectors?.map((connector) => ( - <ListItem - key={connector.name} - connector={connector} - clusterName={clusterName} - /> - ))} - </tbody> - </Table> + <Table + data={connectors || []} + columns={columns} + enableSorting + onRowClick={({ original: { connect, name } }) => + navigate(clusterConnectConnectorPath(clusterName, connect, name)) + } + emptyMessage="No connectors found" + /> ); }; diff --git a/kafka-ui-react-app/src/components/Connect/List/ListItem.tsx b/kafka-ui-react-app/src/components/Connect/List/ListItem.tsx deleted file mode 100644 index 4d10001947b..00000000000 --- a/kafka-ui-react-app/src/components/Connect/List/ListItem.tsx +++ /dev/null @@ -1,98 +0,0 @@ -import React from 'react'; -import { FullConnectorInfo } from 'generated-sources'; -import { clusterConnectConnectorPath, clusterTopicPath } from 'lib/paths'; -import { ClusterName } from 'redux/interfaces'; -import { Link, NavLink } from 'react-router-dom'; -import { Tag } from 'components/common/Tag/Tag.styled'; -import { TableKeyLink } from 'components/common/table/Table/TableKeyLink.styled'; -import getTagColor from 'components/common/Tag/getTagColor'; -import { useDeleteConnector } from 'lib/hooks/api/kafkaConnect'; -import { Dropdown, DropdownItem } from 'components/common/Dropdown'; -import { useConfirm } from 'lib/hooks/useConfirm'; - -import * as S from './List.styled'; - -export interface ListItemProps { - clusterName: ClusterName; - connector: FullConnectorInfo; -} - -const ListItem: React.FC<ListItemProps> = ({ - clusterName, - connector: { - name, - connect, - type, - connectorClass, - topics, - status, - tasksCount, - failedTasksCount, - }, -}) => { - const confirm = useConfirm(); - const deleteMutation = useDeleteConnector({ - clusterName, - connectName: connect, - connectorName: name, - }); - - const handleDelete = () => { - confirm( - <> - Are you sure want to remove <b>{name}</b> connector? - </>, - async () => { - await deleteMutation.mutateAsync(); - } - ); - }; - - const runningTasks = React.useMemo(() => { - if (!tasksCount) return null; - return tasksCount - (failedTasksCount || 0); - }, [tasksCount, failedTasksCount]); - - return ( - <tr> - <TableKeyLink> - <NavLink to={clusterConnectConnectorPath(clusterName, connect, name)}> - {name} - </NavLink> - </TableKeyLink> - <td>{connect}</td> - <td>{type}</td> - <td>{connectorClass}</td> - <td> - <S.TagsWrapper> - {topics?.map((t) => ( - <Tag key={t} color="gray"> - <Link to={clusterTopicPath(clusterName, t)}>{t}</Link> - </Tag> - ))} - </S.TagsWrapper> - </td> - <td> - {status && <Tag color={getTagColor(status.state)}>{status.state}</Tag>} - </td> - <td> - {runningTasks && ( - <span> - {runningTasks} of {tasksCount} - </span> - )} - </td> - <td> - <div> - <Dropdown> - <DropdownItem onClick={handleDelete} danger> - Remove Connector - </DropdownItem> - </Dropdown> - </div> - </td> - </tr> - ); -}; - -export default ListItem; diff --git a/kafka-ui-react-app/src/components/Connect/List/RunningTasksCell.tsx b/kafka-ui-react-app/src/components/Connect/List/RunningTasksCell.tsx new file mode 100644 index 00000000000..4c3293d44c9 --- /dev/null +++ b/kafka-ui-react-app/src/components/Connect/List/RunningTasksCell.tsx @@ -0,0 +1,21 @@ +import React from 'react'; +import { FullConnectorInfo } from 'generated-sources'; +import { CellContext } from '@tanstack/react-table'; + +const RunningTasksCell: React.FC<CellContext<FullConnectorInfo, unknown>> = ({ + row, +}) => { + const { tasksCount, failedTasksCount } = row.original; + + if (!tasksCount) { + return null; + } + + return ( + <> + {tasksCount - (failedTasksCount || 0)} of {tasksCount} + </> + ); +}; + +export default RunningTasksCell; diff --git a/kafka-ui-react-app/src/components/Connect/List/TopicsCell.tsx b/kafka-ui-react-app/src/components/Connect/List/TopicsCell.tsx new file mode 100644 index 00000000000..ee48c2d2d3f --- /dev/null +++ b/kafka-ui-react-app/src/components/Connect/List/TopicsCell.tsx @@ -0,0 +1,45 @@ +import React from 'react'; +import { FullConnectorInfo } from 'generated-sources'; +import { CellContext } from '@tanstack/react-table'; +import { useNavigate } from 'react-router-dom'; +import { Tag } from 'components/common/Tag/Tag.styled'; +import { ClusterNameRoute, clusterTopicPath } from 'lib/paths'; +import useAppParams from 'lib/hooks/useAppParams'; + +import * as S from './List.styled'; + +const TopicsCell: React.FC<CellContext<FullConnectorInfo, unknown>> = ({ + row, +}) => { + const { topics } = row.original; + const { clusterName } = useAppParams<ClusterNameRoute>(); + const navigate = useNavigate(); + + const navigateToTopic = ( + e: React.KeyboardEvent | React.MouseEvent, + topic: string + ) => { + e.preventDefault(); + e.stopPropagation(); + navigate(clusterTopicPath(clusterName, topic)); + }; + + return ( + <S.TagsWrapper> + {topics?.map((t) => ( + <Tag key={t} color="gray"> + <span + role="link" + onClick={(e) => navigateToTopic(e, t)} + onKeyDown={(e) => navigateToTopic(e, t)} + tabIndex={0} + > + {t} + </span> + </Tag> + ))} + </S.TagsWrapper> + ); +}; + +export default TopicsCell; diff --git a/kafka-ui-react-app/src/components/Connect/List/__tests__/List.spec.tsx b/kafka-ui-react-app/src/components/Connect/List/__tests__/List.spec.tsx index a76b8092f25..04a7ba81506 100644 --- a/kafka-ui-react-app/src/components/Connect/List/__tests__/List.spec.tsx +++ b/kafka-ui-react-app/src/components/Connect/List/__tests__/List.spec.tsx @@ -5,49 +5,120 @@ import ClusterContext, { initialValue, } from 'components/contexts/ClusterContext'; import List from 'components/Connect/List/List'; -import { screen } from '@testing-library/react'; +import { act, screen, waitFor } from '@testing-library/react'; +import userEvent from '@testing-library/user-event'; import { render, WithRoute } from 'lib/testHelpers'; -import { clusterConnectorsPath } from 'lib/paths'; -import { useConnectors } from 'lib/hooks/api/kafkaConnect'; - -jest.mock('components/Connect/List/ListItem', () => () => ( - <tr> - <td>List Item</td> - </tr> -)); +import { clusterConnectConnectorPath, clusterConnectorsPath } from 'lib/paths'; +import { useConnectors, useDeleteConnector } from 'lib/hooks/api/kafkaConnect'; + +const mockedUsedNavigate = jest.fn(); +const mockDelete = jest.fn(); + +jest.mock('react-router-dom', () => ({ + ...jest.requireActual('react-router-dom'), + useNavigate: () => mockedUsedNavigate, +})); + jest.mock('lib/hooks/api/kafkaConnect', () => ({ useConnectors: jest.fn(), + useDeleteConnector: jest.fn(), })); const clusterName = 'local'; +const renderComponent = (contextValue: ContextProps = initialValue) => + render( + <ClusterContext.Provider value={contextValue}> + <WithRoute path={clusterConnectorsPath()}> + <List /> + </WithRoute> + </ClusterContext.Provider>, + { initialEntries: [clusterConnectorsPath(clusterName)] } + ); + describe('Connectors List', () => { - const renderComponent = (contextValue: ContextProps = initialValue) => - render( - <ClusterContext.Provider value={contextValue}> - <WithRoute path={clusterConnectorsPath()}> - <List /> - </WithRoute> - </ClusterContext.Provider>, - { initialEntries: [clusterConnectorsPath(clusterName)] } - ); - - it('renders empty connectors Table', async () => { - (useConnectors as jest.Mock).mockImplementation(() => ({ - data: [], - })); - - await renderComponent(); - expect(screen.getByRole('table')).toBeInTheDocument(); - expect(screen.getByText('No connectors found')).toBeInTheDocument(); + describe('when the connectors are loaded', () => { + beforeEach(() => { + (useConnectors as jest.Mock).mockImplementation(() => ({ + data: connectors, + })); + }); + + it('renders', async () => { + renderComponent(); + expect(screen.getByRole('table')).toBeInTheDocument(); + expect(screen.getAllByRole('row').length).toEqual(3); + }); + + it('opens broker when row clicked', async () => { + renderComponent(); + await act(() => { + userEvent.click( + screen.getByRole('row', { + name: 'hdfs-source-connector first SOURCE FileStreamSource a b c RUNNING 2 of 2', + }) + ); + }); + await waitFor(() => + expect(mockedUsedNavigate).toBeCalledWith( + clusterConnectConnectorPath( + clusterName, + 'first', + 'hdfs-source-connector' + ) + ) + ); + }); }); - it('renders connectors Table', async () => { - (useConnectors as jest.Mock).mockImplementation(() => ({ - data: connectors, - })); - await renderComponent(); - expect(screen.getByRole('table')).toBeInTheDocument(); - expect(screen.getAllByText('List Item').length).toEqual(2); + describe('when table is empty', () => { + beforeEach(() => { + (useConnectors as jest.Mock).mockImplementation(() => ({ + data: [], + })); + }); + + it('renders empty table', async () => { + renderComponent(); + expect(screen.getByRole('table')).toBeInTheDocument(); + expect( + screen.getByRole('row', { name: 'No connectors found' }) + ).toBeInTheDocument(); + }); + }); + + describe('when remove connector modal is open', () => { + beforeEach(() => { + (useConnectors as jest.Mock).mockImplementation(() => ({ + data: connectors, + })); + (useDeleteConnector as jest.Mock).mockImplementation(() => ({ + mutateAsync: mockDelete, + })); + }); + + it('calls removeConnector on confirm', async () => { + renderComponent(); + const removeButton = screen.getAllByText('Remove Connector')[0]; + await waitFor(() => userEvent.click(removeButton)); + + const submitButton = screen.getAllByRole('button', { + name: 'Confirm', + })[0]; + await act(() => userEvent.click(submitButton)); + expect(mockDelete).toHaveBeenCalledWith(); + }); + + it('closes the modal when cancel button is clicked', async () => { + renderComponent(); + const removeButton = screen.getAllByText('Remove Connector')[0]; + await waitFor(() => userEvent.click(removeButton)); + + const cancelButton = screen.getAllByRole('button', { + name: 'Cancel', + })[0]; + await waitFor(() => userEvent.click(cancelButton)); + expect(cancelButton).not.toBeInTheDocument(); + }); }); }); diff --git a/kafka-ui-react-app/src/components/Connect/List/__tests__/ListItem.spec.tsx b/kafka-ui-react-app/src/components/Connect/List/__tests__/ListItem.spec.tsx deleted file mode 100644 index 715d0aeb4d3..00000000000 --- a/kafka-ui-react-app/src/components/Connect/List/__tests__/ListItem.spec.tsx +++ /dev/null @@ -1,53 +0,0 @@ -import React from 'react'; -import { connectors } from 'lib/fixtures/kafkaConnect'; -import ListItem, { ListItemProps } from 'components/Connect/List/ListItem'; -import { screen } from '@testing-library/react'; -import { render } from 'lib/testHelpers'; - -describe('Connectors ListItem', () => { - const connector = connectors[0]; - const setupWrapper = (props: Partial<ListItemProps> = {}) => ( - <table> - <tbody> - <ListItem clusterName="local" connector={connector} {...props} /> - </tbody> - </table> - ); - - it('renders item', () => { - render(setupWrapper()); - expect(screen.getAllByRole('cell')[6]).toHaveTextContent('2 of 2'); - }); - - it('topics tags are sorted', () => { - render(setupWrapper()); - const getLink = screen.getAllByRole('link'); - expect(getLink[1]).toHaveTextContent('a'); - expect(getLink[2]).toHaveTextContent('b'); - expect(getLink[3]).toHaveTextContent('c'); - }); - - it('renders item with failed tasks', () => { - render( - setupWrapper({ - connector: { - ...connector, - failedTasksCount: 1, - }, - }) - ); - expect(screen.getAllByRole('cell')[6]).toHaveTextContent('1 of 2'); - }); - - it('does not render info about tasks if taksCount is undefined', () => { - render( - setupWrapper({ - connector: { - ...connector, - tasksCount: undefined, - }, - }) - ); - expect(screen.getAllByRole('cell')[6]).toHaveTextContent(''); - }); -});
null
train
val
2022-10-07T15:04:21
"2022-05-31T13:04:31Z"
Haarolean
train
provectus/kafka-ui/2325_2689
provectus/kafka-ui
provectus/kafka-ui/2325
provectus/kafka-ui/2689
[ "keyword_pr_to_issue", "timestamp(timedelta=119.0, similarity=0.9020096887373107)" ]
80eb2dccfea9b15b64e927493fbbc5158549c935
7d5b7de992da9ddb1a018a6a61db66c4dcb204aa
[]
[ "Can we use existing `NewTable/TagCell` component?", "```suggestion\r\nconst TopicsCell: React.FC<CellContext<FullConnectorInfo, unknown>> = ({\r\n```", "useMemo is redundant here", "```suggestion\r\n if (!tasksCount) {\r\n return null;\r\n }\r\n \r\n return (tasksCount ? \r\n <>\r\n {tasksCount - (failedTasksCount || 0)} of {tasksCount}\r\n </>\r\n );\r\n```", "Pls use styled components attr to change link color. Or update Tag styles to show link correctly" ]
"2022-10-01T02:20:41Z"
[ "type/enhancement", "good first issue", "scope/frontend", "status/accepted", "hacktoberfest" ]
Make connectors table rows clickable
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java", "kafka-ui-react-app/src/components/Connect/List/List.styled.ts", "kafka-ui-react-app/src/components/Connect/List/List.tsx", "kafka-ui-react-app/src/components/Connect/List/ListItem.tsx", "kafka-ui-react-app/src/components/Connect/List/__tests__/List.spec.tsx", "kafka-ui-react-app/src/components/Connect/List/__tests__/ListItem.spec.tsx" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java", "kafka-ui-react-app/src/components/Connect/List/ActionsCell.tsx", "kafka-ui-react-app/src/components/Connect/List/List.styled.ts", "kafka-ui-react-app/src/components/Connect/List/List.tsx", "kafka-ui-react-app/src/components/Connect/List/RunningTasksCell.tsx", "kafka-ui-react-app/src/components/Connect/List/TopicsCell.tsx", "kafka-ui-react-app/src/components/Connect/List/__tests__/List.spec.tsx" ]
[]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java index b5986c7919e..89a73ff6931 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java @@ -28,14 +28,14 @@ public ConnectorCreateForm clickCreateConnectorButton() { @Step public KafkaConnectList openConnector(String connectorName) { - $(By.linkText(connectorName)).click(); + $x("//tbody//td[1][text()='" + connectorName + "']").shouldBe(Condition.enabled).click(); return this; } @Step public boolean isConnectorVisible(String connectorName) { $(By.xpath("//table")).shouldBe(Condition.visible); - return isVisible($x("//tbody//td[1]//a[text()='" + connectorName + "']")); + return isVisible($x("//tbody//td[1][text()='" + connectorName + "']")); } @Step diff --git a/kafka-ui-react-app/src/components/Connect/List/ActionsCell.tsx b/kafka-ui-react-app/src/components/Connect/List/ActionsCell.tsx new file mode 100644 index 00000000000..30b3df8a56d --- /dev/null +++ b/kafka-ui-react-app/src/components/Connect/List/ActionsCell.tsx @@ -0,0 +1,43 @@ +import React from 'react'; +import { FullConnectorInfo } from 'generated-sources'; +import { CellContext } from '@tanstack/react-table'; +import { ClusterNameRoute } from 'lib/paths'; +import useAppParams from 'lib/hooks/useAppParams'; +import { Dropdown, DropdownItem } from 'components/common/Dropdown'; +import { useDeleteConnector } from 'lib/hooks/api/kafkaConnect'; +import { useConfirm } from 'lib/hooks/useConfirm'; + +const ActionsCell: React.FC<CellContext<FullConnectorInfo, unknown>> = ({ + row, +}) => { + const { connect, name } = row.original; + + const { clusterName } = useAppParams<ClusterNameRoute>(); + + const confirm = useConfirm(); + const deleteMutation = useDeleteConnector({ + clusterName, + connectName: connect, + connectorName: name, + }); + + const handleDelete = () => { + confirm( + <> + Are you sure want to remove <b>{name}</b> connector? + </>, + async () => { + await deleteMutation.mutateAsync(); + } + ); + }; + return ( + <Dropdown> + <DropdownItem onClick={handleDelete} danger> + Remove Connector + </DropdownItem> + </Dropdown> + ); +}; + +export default ActionsCell; diff --git a/kafka-ui-react-app/src/components/Connect/List/List.styled.ts b/kafka-ui-react-app/src/components/Connect/List/List.styled.ts index f0e2631d2c1..799915fcb11 100644 --- a/kafka-ui-react-app/src/components/Connect/List/List.styled.ts +++ b/kafka-ui-react-app/src/components/Connect/List/List.styled.ts @@ -3,4 +3,10 @@ import styled from 'styled-components'; export const TagsWrapper = styled.div` display: flex; flex-wrap: wrap; + span { + color: rgb(76, 76, 255) !important; + &:hover { + color: rgb(23, 23, 207) !important; + } + } `; diff --git a/kafka-ui-react-app/src/components/Connect/List/List.tsx b/kafka-ui-react-app/src/components/Connect/List/List.tsx index 75a782fa0a6..b5935e7bab2 100644 --- a/kafka-ui-react-app/src/components/Connect/List/List.tsx +++ b/kafka-ui-react-app/src/components/Connect/List/List.tsx @@ -1,14 +1,18 @@ import React from 'react'; import useAppParams from 'lib/hooks/useAppParams'; -import { ClusterNameRoute } from 'lib/paths'; -import { Table } from 'components/common/table/Table/Table.styled'; -import TableHeaderCell from 'components/common/table/TableHeaderCell/TableHeaderCell'; +import { clusterConnectConnectorPath, ClusterNameRoute } from 'lib/paths'; +import Table, { TagCell } from 'components/common/NewTable'; +import { FullConnectorInfo } from 'generated-sources'; import { useConnectors } from 'lib/hooks/api/kafkaConnect'; -import { useSearchParams } from 'react-router-dom'; +import { ColumnDef } from '@tanstack/react-table'; +import { useNavigate, useSearchParams } from 'react-router-dom'; -import ListItem from './ListItem'; +import ActionsCell from './ActionsCell'; +import TopicsCell from './TopicsCell'; +import RunningTasksCell from './RunningTasksCell'; const List: React.FC = () => { + const navigate = useNavigate(); const { clusterName } = useAppParams<ClusterNameRoute>(); const [searchParams] = useSearchParams(); const { data: connectors } = useConnectors( @@ -16,35 +20,30 @@ const List: React.FC = () => { searchParams.get('q') || '' ); + const columns = React.useMemo<ColumnDef<FullConnectorInfo>[]>( + () => [ + { header: 'Name', accessorKey: 'name' }, + { header: 'Connect', accessorKey: 'connect' }, + { header: 'Type', accessorKey: 'type' }, + { header: 'Plugin', accessorKey: 'connectorClass' }, + { header: 'Topics', cell: TopicsCell }, + { header: 'Status', accessorKey: 'status.state', cell: TagCell }, + { header: 'Running Tasks', cell: RunningTasksCell }, + { header: '', id: 'action', cell: ActionsCell }, + ], + [] + ); + return ( - <Table isFullwidth> - <thead> - <tr> - <TableHeaderCell title="Name" /> - <TableHeaderCell title="Connect" /> - <TableHeaderCell title="Type" /> - <TableHeaderCell title="Plugin" /> - <TableHeaderCell title="Topics" /> - <TableHeaderCell title="Status" /> - <TableHeaderCell title="Running Tasks" /> - <TableHeaderCell> </TableHeaderCell> - </tr> - </thead> - <tbody> - {(!connectors || connectors.length) === 0 && ( - <tr> - <td colSpan={10}>No connectors found</td> - </tr> - )} - {connectors?.map((connector) => ( - <ListItem - key={connector.name} - connector={connector} - clusterName={clusterName} - /> - ))} - </tbody> - </Table> + <Table + data={connectors || []} + columns={columns} + enableSorting + onRowClick={({ original: { connect, name } }) => + navigate(clusterConnectConnectorPath(clusterName, connect, name)) + } + emptyMessage="No connectors found" + /> ); }; diff --git a/kafka-ui-react-app/src/components/Connect/List/ListItem.tsx b/kafka-ui-react-app/src/components/Connect/List/ListItem.tsx deleted file mode 100644 index 4d10001947b..00000000000 --- a/kafka-ui-react-app/src/components/Connect/List/ListItem.tsx +++ /dev/null @@ -1,98 +0,0 @@ -import React from 'react'; -import { FullConnectorInfo } from 'generated-sources'; -import { clusterConnectConnectorPath, clusterTopicPath } from 'lib/paths'; -import { ClusterName } from 'redux/interfaces'; -import { Link, NavLink } from 'react-router-dom'; -import { Tag } from 'components/common/Tag/Tag.styled'; -import { TableKeyLink } from 'components/common/table/Table/TableKeyLink.styled'; -import getTagColor from 'components/common/Tag/getTagColor'; -import { useDeleteConnector } from 'lib/hooks/api/kafkaConnect'; -import { Dropdown, DropdownItem } from 'components/common/Dropdown'; -import { useConfirm } from 'lib/hooks/useConfirm'; - -import * as S from './List.styled'; - -export interface ListItemProps { - clusterName: ClusterName; - connector: FullConnectorInfo; -} - -const ListItem: React.FC<ListItemProps> = ({ - clusterName, - connector: { - name, - connect, - type, - connectorClass, - topics, - status, - tasksCount, - failedTasksCount, - }, -}) => { - const confirm = useConfirm(); - const deleteMutation = useDeleteConnector({ - clusterName, - connectName: connect, - connectorName: name, - }); - - const handleDelete = () => { - confirm( - <> - Are you sure want to remove <b>{name}</b> connector? - </>, - async () => { - await deleteMutation.mutateAsync(); - } - ); - }; - - const runningTasks = React.useMemo(() => { - if (!tasksCount) return null; - return tasksCount - (failedTasksCount || 0); - }, [tasksCount, failedTasksCount]); - - return ( - <tr> - <TableKeyLink> - <NavLink to={clusterConnectConnectorPath(clusterName, connect, name)}> - {name} - </NavLink> - </TableKeyLink> - <td>{connect}</td> - <td>{type}</td> - <td>{connectorClass}</td> - <td> - <S.TagsWrapper> - {topics?.map((t) => ( - <Tag key={t} color="gray"> - <Link to={clusterTopicPath(clusterName, t)}>{t}</Link> - </Tag> - ))} - </S.TagsWrapper> - </td> - <td> - {status && <Tag color={getTagColor(status.state)}>{status.state}</Tag>} - </td> - <td> - {runningTasks && ( - <span> - {runningTasks} of {tasksCount} - </span> - )} - </td> - <td> - <div> - <Dropdown> - <DropdownItem onClick={handleDelete} danger> - Remove Connector - </DropdownItem> - </Dropdown> - </div> - </td> - </tr> - ); -}; - -export default ListItem; diff --git a/kafka-ui-react-app/src/components/Connect/List/RunningTasksCell.tsx b/kafka-ui-react-app/src/components/Connect/List/RunningTasksCell.tsx new file mode 100644 index 00000000000..4c3293d44c9 --- /dev/null +++ b/kafka-ui-react-app/src/components/Connect/List/RunningTasksCell.tsx @@ -0,0 +1,21 @@ +import React from 'react'; +import { FullConnectorInfo } from 'generated-sources'; +import { CellContext } from '@tanstack/react-table'; + +const RunningTasksCell: React.FC<CellContext<FullConnectorInfo, unknown>> = ({ + row, +}) => { + const { tasksCount, failedTasksCount } = row.original; + + if (!tasksCount) { + return null; + } + + return ( + <> + {tasksCount - (failedTasksCount || 0)} of {tasksCount} + </> + ); +}; + +export default RunningTasksCell; diff --git a/kafka-ui-react-app/src/components/Connect/List/TopicsCell.tsx b/kafka-ui-react-app/src/components/Connect/List/TopicsCell.tsx new file mode 100644 index 00000000000..ee48c2d2d3f --- /dev/null +++ b/kafka-ui-react-app/src/components/Connect/List/TopicsCell.tsx @@ -0,0 +1,45 @@ +import React from 'react'; +import { FullConnectorInfo } from 'generated-sources'; +import { CellContext } from '@tanstack/react-table'; +import { useNavigate } from 'react-router-dom'; +import { Tag } from 'components/common/Tag/Tag.styled'; +import { ClusterNameRoute, clusterTopicPath } from 'lib/paths'; +import useAppParams from 'lib/hooks/useAppParams'; + +import * as S from './List.styled'; + +const TopicsCell: React.FC<CellContext<FullConnectorInfo, unknown>> = ({ + row, +}) => { + const { topics } = row.original; + const { clusterName } = useAppParams<ClusterNameRoute>(); + const navigate = useNavigate(); + + const navigateToTopic = ( + e: React.KeyboardEvent | React.MouseEvent, + topic: string + ) => { + e.preventDefault(); + e.stopPropagation(); + navigate(clusterTopicPath(clusterName, topic)); + }; + + return ( + <S.TagsWrapper> + {topics?.map((t) => ( + <Tag key={t} color="gray"> + <span + role="link" + onClick={(e) => navigateToTopic(e, t)} + onKeyDown={(e) => navigateToTopic(e, t)} + tabIndex={0} + > + {t} + </span> + </Tag> + ))} + </S.TagsWrapper> + ); +}; + +export default TopicsCell; diff --git a/kafka-ui-react-app/src/components/Connect/List/__tests__/List.spec.tsx b/kafka-ui-react-app/src/components/Connect/List/__tests__/List.spec.tsx index a76b8092f25..04a7ba81506 100644 --- a/kafka-ui-react-app/src/components/Connect/List/__tests__/List.spec.tsx +++ b/kafka-ui-react-app/src/components/Connect/List/__tests__/List.spec.tsx @@ -5,49 +5,120 @@ import ClusterContext, { initialValue, } from 'components/contexts/ClusterContext'; import List from 'components/Connect/List/List'; -import { screen } from '@testing-library/react'; +import { act, screen, waitFor } from '@testing-library/react'; +import userEvent from '@testing-library/user-event'; import { render, WithRoute } from 'lib/testHelpers'; -import { clusterConnectorsPath } from 'lib/paths'; -import { useConnectors } from 'lib/hooks/api/kafkaConnect'; - -jest.mock('components/Connect/List/ListItem', () => () => ( - <tr> - <td>List Item</td> - </tr> -)); +import { clusterConnectConnectorPath, clusterConnectorsPath } from 'lib/paths'; +import { useConnectors, useDeleteConnector } from 'lib/hooks/api/kafkaConnect'; + +const mockedUsedNavigate = jest.fn(); +const mockDelete = jest.fn(); + +jest.mock('react-router-dom', () => ({ + ...jest.requireActual('react-router-dom'), + useNavigate: () => mockedUsedNavigate, +})); + jest.mock('lib/hooks/api/kafkaConnect', () => ({ useConnectors: jest.fn(), + useDeleteConnector: jest.fn(), })); const clusterName = 'local'; +const renderComponent = (contextValue: ContextProps = initialValue) => + render( + <ClusterContext.Provider value={contextValue}> + <WithRoute path={clusterConnectorsPath()}> + <List /> + </WithRoute> + </ClusterContext.Provider>, + { initialEntries: [clusterConnectorsPath(clusterName)] } + ); + describe('Connectors List', () => { - const renderComponent = (contextValue: ContextProps = initialValue) => - render( - <ClusterContext.Provider value={contextValue}> - <WithRoute path={clusterConnectorsPath()}> - <List /> - </WithRoute> - </ClusterContext.Provider>, - { initialEntries: [clusterConnectorsPath(clusterName)] } - ); - - it('renders empty connectors Table', async () => { - (useConnectors as jest.Mock).mockImplementation(() => ({ - data: [], - })); - - await renderComponent(); - expect(screen.getByRole('table')).toBeInTheDocument(); - expect(screen.getByText('No connectors found')).toBeInTheDocument(); + describe('when the connectors are loaded', () => { + beforeEach(() => { + (useConnectors as jest.Mock).mockImplementation(() => ({ + data: connectors, + })); + }); + + it('renders', async () => { + renderComponent(); + expect(screen.getByRole('table')).toBeInTheDocument(); + expect(screen.getAllByRole('row').length).toEqual(3); + }); + + it('opens broker when row clicked', async () => { + renderComponent(); + await act(() => { + userEvent.click( + screen.getByRole('row', { + name: 'hdfs-source-connector first SOURCE FileStreamSource a b c RUNNING 2 of 2', + }) + ); + }); + await waitFor(() => + expect(mockedUsedNavigate).toBeCalledWith( + clusterConnectConnectorPath( + clusterName, + 'first', + 'hdfs-source-connector' + ) + ) + ); + }); }); - it('renders connectors Table', async () => { - (useConnectors as jest.Mock).mockImplementation(() => ({ - data: connectors, - })); - await renderComponent(); - expect(screen.getByRole('table')).toBeInTheDocument(); - expect(screen.getAllByText('List Item').length).toEqual(2); + describe('when table is empty', () => { + beforeEach(() => { + (useConnectors as jest.Mock).mockImplementation(() => ({ + data: [], + })); + }); + + it('renders empty table', async () => { + renderComponent(); + expect(screen.getByRole('table')).toBeInTheDocument(); + expect( + screen.getByRole('row', { name: 'No connectors found' }) + ).toBeInTheDocument(); + }); + }); + + describe('when remove connector modal is open', () => { + beforeEach(() => { + (useConnectors as jest.Mock).mockImplementation(() => ({ + data: connectors, + })); + (useDeleteConnector as jest.Mock).mockImplementation(() => ({ + mutateAsync: mockDelete, + })); + }); + + it('calls removeConnector on confirm', async () => { + renderComponent(); + const removeButton = screen.getAllByText('Remove Connector')[0]; + await waitFor(() => userEvent.click(removeButton)); + + const submitButton = screen.getAllByRole('button', { + name: 'Confirm', + })[0]; + await act(() => userEvent.click(submitButton)); + expect(mockDelete).toHaveBeenCalledWith(); + }); + + it('closes the modal when cancel button is clicked', async () => { + renderComponent(); + const removeButton = screen.getAllByText('Remove Connector')[0]; + await waitFor(() => userEvent.click(removeButton)); + + const cancelButton = screen.getAllByRole('button', { + name: 'Cancel', + })[0]; + await waitFor(() => userEvent.click(cancelButton)); + expect(cancelButton).not.toBeInTheDocument(); + }); }); }); diff --git a/kafka-ui-react-app/src/components/Connect/List/__tests__/ListItem.spec.tsx b/kafka-ui-react-app/src/components/Connect/List/__tests__/ListItem.spec.tsx deleted file mode 100644 index 715d0aeb4d3..00000000000 --- a/kafka-ui-react-app/src/components/Connect/List/__tests__/ListItem.spec.tsx +++ /dev/null @@ -1,53 +0,0 @@ -import React from 'react'; -import { connectors } from 'lib/fixtures/kafkaConnect'; -import ListItem, { ListItemProps } from 'components/Connect/List/ListItem'; -import { screen } from '@testing-library/react'; -import { render } from 'lib/testHelpers'; - -describe('Connectors ListItem', () => { - const connector = connectors[0]; - const setupWrapper = (props: Partial<ListItemProps> = {}) => ( - <table> - <tbody> - <ListItem clusterName="local" connector={connector} {...props} /> - </tbody> - </table> - ); - - it('renders item', () => { - render(setupWrapper()); - expect(screen.getAllByRole('cell')[6]).toHaveTextContent('2 of 2'); - }); - - it('topics tags are sorted', () => { - render(setupWrapper()); - const getLink = screen.getAllByRole('link'); - expect(getLink[1]).toHaveTextContent('a'); - expect(getLink[2]).toHaveTextContent('b'); - expect(getLink[3]).toHaveTextContent('c'); - }); - - it('renders item with failed tasks', () => { - render( - setupWrapper({ - connector: { - ...connector, - failedTasksCount: 1, - }, - }) - ); - expect(screen.getAllByRole('cell')[6]).toHaveTextContent('1 of 2'); - }); - - it('does not render info about tasks if taksCount is undefined', () => { - render( - setupWrapper({ - connector: { - ...connector, - tasksCount: undefined, - }, - }) - ); - expect(screen.getAllByRole('cell')[6]).toHaveTextContent(''); - }); -});
null
train
val
2022-10-07T15:04:21
"2022-07-22T22:23:14Z"
Haarolean
train
provectus/kafka-ui/2634_2690
provectus/kafka-ui
provectus/kafka-ui/2634
provectus/kafka-ui/2690
[ "keyword_pr_to_issue", "timestamp(timedelta=1.0, similarity=0.9034101006360602)" ]
55c8d6ac2d4bdfb9721c79d4894732c86e0380d9
e3aeb0cac8d33d3dd9b9b7232e28fd7c613c0bef
[ "Hello, could I please be assigned to this issue?", "> Hello, could I please be assigned to this issue?\r\n\r\nsure, go ahead :)", "Is this issue still active?", "Depends on the definition β€œactive”\n\n> On 29 Sep 2022, at 17:00, Aditya Bhattad ***@***.***> wrote:\n> \n> ο»Ώ\n> Is this issue still active?\n> \n> β€”\n> Reply to this email directly, view it on GitHub, or unsubscribe.\n> You are receiving this because you commented.\n", "Hello,\n\nSorry for the confusionβ€” I chose another issue for my project, so it is up\nfor grabs now.\n\nOn Thu, Sep 29, 2022 at 9:00 AM Aditya Bhattad ***@***.***>\nwrote:\n\n> Is this issue still active?\n>\n> β€”\n> Reply to this email directly, view it on GitHub\n> <https://github.com/provectus/kafka-ui/issues/2634#issuecomment-1262243988>,\n> or unsubscribe\n> <https://github.com/notifications/unsubscribe-auth/AWT2UK3QPMLIDHH3ANCWMTLWAWHGPANCNFSM6AAAAAAQTGPN44>\n> .\n> You are receiving this because you were assigned.Message ID:\n> ***@***.***>\n>\n-- \n*Caroline Kraus*\n", "Can I work on this?", "Sure! Go ahead :)\n\n> On 29 Sep 2022, at 21:44, Aditya Bhattad ***@***.***> wrote:\n> \n> ο»Ώ\n> Can I work on this?\n> \n> β€”\n> Reply to this email directly, view it on GitHub, or unsubscribe.\n> You are receiving this because you commented.\n", "Hey I am trying to set up `kafka-ui-react-app` for local development, I try to follow set guide on README.md of the `kafka-ui-react-app` but I cannot understand these two steps :\r\nGenerate API clients from OpenAPI document\r\n```sh\r\npnpm gen:sources\r\n```\r\nand \r\nCreate or update existing `.env.local` file with\r\n```\r\nVITE_DEV_PROXY= https://api.server # your API server\r\n```\r\nwhat does `pnpm gen:sources` do, is setting this: `VITE_DEV_PROXY= https://api.server` is necessary for local development? ", "@adityabhattad2021 \r\n1. `gen:sources` does generate classes from API declaration (kafka-ui-api.yaml) via openapi-generator.\r\n2. It's the URL for the backend. You can run kafka-ui docker container and expose the 8080 port, and set dev proxy to http://localhost:8080. \r\nLet me know how it goes.", "hey `pnpm gen:sources` gives error on my device \r\n![Screenshot 2022-09-30 174856](https://user-images.githubusercontent.com/93488388/193268014-437ba64d-1abd-403a-8303-11556c5e9e0f.png)\r\ndoes it needs java setup on the device?\r\nand I do not have docker setup currently.\r\nI made changes anyways, will make a pull request tomorrow.\r\nActually I am a newbie just trying to get into open source.\r\nThanks for your time.", "> hey `pnpm gen:sources` gives error on my device ![Screenshot 2022-09-30 174856](https://user-images.githubusercontent.com/93488388/193268014-437ba64d-1abd-403a-8303-11556c5e9e0f.png) does it needs java setup on the device? and I do not have docker setup currently. I made changes anyways, will make a pull request tomorrow. Actually I am a newbie just trying to get into open source. Thanks for your time.\r\n\r\nyep you do :)\r\nYou can build the app locally with maven, but it'd be easier with docker, you won't have to build the app at all, just the frontend part.\r\nSure, any contributions are welcome :)" ]
[]
"2022-10-01T02:59:12Z"
[ "good first issue", "scope/frontend", "status/accepted", "type/chore", "hacktoberfest" ]
[UI] Delete "Num Of Topics" column from Topic->Consumers page
Where: Topics -> select topic -> Consumers Todo: 1. Remove "Num of Topics" columns 2. Rename "Num of Members" column to "Active consumers" <img width="1511" alt="Screenshot 2022-09-22 at 19 48 54" src="https://user-images.githubusercontent.com/702205/191796514-46e1b853-ea21-49c8-b932-dafc45168390.png">
[ "kafka-ui-react-app/src/components/Topics/Topic/ConsumerGroups/TopicConsumerGroups.tsx" ]
[ "kafka-ui-react-app/src/components/Topics/Topic/ConsumerGroups/TopicConsumerGroups.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/Topic/ConsumerGroups/TopicConsumerGroups.tsx b/kafka-ui-react-app/src/components/Topics/Topic/ConsumerGroups/TopicConsumerGroups.tsx index d1d6c4ece2c..0fc72fd82e8 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/ConsumerGroups/TopicConsumerGroups.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/ConsumerGroups/TopicConsumerGroups.tsx @@ -21,8 +21,7 @@ const TopicConsumerGroups: React.FC = () => { <thead> <tr> <TableHeaderCell title="Consumer Group ID" /> - <TableHeaderCell title="Num Of Members" /> - <TableHeaderCell title="Num Of Topics" /> + <TableHeaderCell title="Active Consumers" /> <TableHeaderCell title="Messages Behind" /> <TableHeaderCell title="Coordinator" /> <TableHeaderCell title="State" /> @@ -41,7 +40,6 @@ const TopicConsumerGroups: React.FC = () => { </Link> </TableKeyLink> <td>{consumer.members}</td> - <td>{consumer.topics}</td> <td>{consumer.messagesBehind}</td> <td>{consumer.coordinator?.id}</td> <td>
null
train
val
2022-10-02T09:34:13
"2022-09-22T16:02:41Z"
iliax
train
provectus/kafka-ui/2674_2691
provectus/kafka-ui
provectus/kafka-ui/2674
provectus/kafka-ui/2691
[ "keyword_pr_to_issue" ]
0818bb0144ca4b2f3b23a0ab9e374fa8bf727901
20543fc2f3830968d8cd9a6a3956f31157c36b6d
[ "@Haarolean it looks almost according the way you suggested. However the overall progress percentage and progress bar look a bit strange now:\r\n![image](https://user-images.githubusercontent.com/112083452/200502621-64bee7cb-3947-4afa-9753-b741b14716e3.png)\r\n\r\nSuggestion: Show the percentage above the progress bar, using a bigger font (most important info in the block, while the rest is additional info).\r\nAnd the progress bar width to be matched to block width and it's alignment to the center. So it would be more consistent.\r\n\r\nShould it be created as a separate improvement task or fixed here while we are at it anyway?\r\n\r\nCC @armenuikafka ", "![pct_statistics](https://github.com/provectus/kafka-ui/assets/894746/059ec5eb-c6ca-4d87-9f9e-79e95d1f3b8e)\r\nWere you looking for something like this? @BulatKha \r\nI can submit a PR, can you assign this to me? @Haarolean ", "@pradyumnad that's better, yes", "@Haarolean PR submitted https://github.com/provectus/kafka-ui/pull/4123\r\n", "@Haarolean @pradyumnad just a minor comment to fix the alignment of the progress bar, as it looks a bit sideways. Centered and stretched from the left edge of the text to the right border of the second column would look better.\r\nLooks fine otherwise!", "@BulatKha yes, I was just fiddling with Pixelmator to kind of get the idea. The code was done to fix it." ]
[ "> Display time passed since the start\r\n\r\nI see that you show time passed since component mount not since analyzing start", "pls remove it", "pls add tests for `calculateTimer` ", "what are you testing here?", "should be wrapped to `it`" ]
"2022-10-01T06:02:32Z"
[ "type/enhancement", "good first issue", "scope/frontend", "status/accepted", "status/design-review" ]
Topic analysis improvements
<img width="351" alt="image" src="https://user-images.githubusercontent.com/1494347/193237442-2e91ab1c-cc49-40e1-b9de-935d381f08a0.png"> - Display percentage in % - Display time passed since the start - Split "scanned messages" to two lines: "scanned messages" and "scanned bytes"
[ "kafka-ui-react-app/src/components/Topics/Topic/Statistics/Metrics.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Statistics/Statistics.styles.ts", "kafka-ui-react-app/src/components/Topics/Topic/Statistics/__test__/Metrics.spec.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Statistics/__test__/Statistics.spec.tsx", "kafka-ui-react-app/src/lib/dateTimeHelpers.ts" ]
[ "kafka-ui-react-app/src/components/Topics/Topic/Statistics/Metrics.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Statistics/Statistics.styles.ts", "kafka-ui-react-app/src/components/Topics/Topic/Statistics/__test__/Metrics.spec.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Statistics/__test__/Statistics.spec.tsx", "kafka-ui-react-app/src/lib/__test__/dateTimeHelpers.spec.ts", "kafka-ui-react-app/src/lib/dateTimeHelpers.ts" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Statistics/Metrics.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Statistics/Metrics.tsx index 34feb01b72e..40abb14d705 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Statistics/Metrics.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Statistics/Metrics.tsx @@ -1,4 +1,4 @@ -import React from 'react'; +import React, { useEffect, useState } from 'react'; import { useAnalyzeTopic, useCancelTopicAnalysis, @@ -15,6 +15,7 @@ import { } from 'components/common/PropertiesList/PropertiesList.styled'; import BytesFormatted from 'components/common/BytesFormatted/BytesFormatted'; import { useTimeFormat } from 'lib/hooks/useTimeFormat'; +import { calculateTimer } from 'lib/dateTimeHelpers'; import * as S from './Statistics.styles'; import Total from './Indicators/Total'; @@ -25,13 +26,14 @@ const Metrics: React.FC = () => { const formatTimestamp = useTimeFormat(); const params = useAppParams<RouteParamsClusterTopic>(); - const [isAnalyzing, setIsAnalyzing] = React.useState(true); + + const [isAnalyzing, setIsAnalyzing] = useState(true); const analyzeTopic = useAnalyzeTopic(params); const cancelTopicAnalysis = useCancelTopicAnalysis(params); const { data } = useTopicAnalysis(params, isAnalyzing); - React.useEffect(() => { + useEffect(() => { if (data && !data.progress) { setIsAnalyzing(false); } @@ -44,7 +46,10 @@ const Metrics: React.FC = () => { if (data.progress) { return ( <S.ProgressContainer> - <ProgressBar completed={data.progress.completenessPercent || 0} /> + <S.ProgressBarWrapper> + <ProgressBar completed={data.progress.completenessPercent || 0} /> + <span> {Math.floor(data.progress.completenessPercent || 0)} %</span> + </S.ProgressBarWrapper> <Button onClick={async () => { await cancelTopicAnalysis.mutateAsync(); @@ -58,9 +63,12 @@ const Metrics: React.FC = () => { <List> <Label>Started at</Label> <span>{formatTimestamp(data.progress.startedAt, 'hh:mm:ss a')}</span> + <Label>Passed since start</Label> + <span>{calculateTimer(data.progress.startedAt as number)}</span> <Label>Scanned messages</Label> + <span>{data.progress.msgsScanned}</span> + <Label>Scanned size</Label> <span> - {data.progress.msgsScanned} /{' '} <BytesFormatted value={data.progress.bytesScanned} /> </span> </List> @@ -90,7 +98,6 @@ const Metrics: React.FC = () => { Restart Analysis </Button> </S.ActionsBar> - <Informers.Wrapper> <Total {...totalStats} /> {totalStats.keySize && ( diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Statistics/Statistics.styles.ts b/kafka-ui-react-app/src/components/Topics/Topic/Statistics/Statistics.styles.ts index 472562eef9d..06a4ae9c9d8 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Statistics/Statistics.styles.ts +++ b/kafka-ui-react-app/src/components/Topics/Topic/Statistics/Statistics.styles.ts @@ -1,7 +1,4 @@ -import { - Label, - List, -} from 'components/common/PropertiesList/PropertiesList.styled'; +import { List } from 'components/common/PropertiesList/PropertiesList.styled'; import styled from 'styled-components'; export const ProgressContainer = styled.div` @@ -16,10 +13,6 @@ export const ProgressContainer = styled.div` ${List} { opacity: 0.5; - - ${Label} { - text-align: right; - } } `; @@ -42,3 +35,10 @@ export const PartitionInfo = styled.div` grid-template-columns: repeat(auto-fit, minmax(300px, 1fr)); column-gap: 24px; `; + +export const ProgressBarWrapper = styled.div` + display: flex; + justify-content: space-between; + align-items: center; + width: 280px; +`; diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Statistics/__test__/Metrics.spec.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Statistics/__test__/Metrics.spec.tsx index 2f27802a002..299a53f4a7b 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Statistics/__test__/Metrics.spec.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Statistics/__test__/Metrics.spec.tsx @@ -34,6 +34,7 @@ describe('Metrics', () => { describe('when analysis is in progress', () => { const cancelMock = jest.fn(); + beforeEach(() => { (useCancelTopicAnalysis as jest.Mock).mockImplementation(() => ({ mutateAsync: cancelMock, @@ -50,10 +51,10 @@ describe('Metrics', () => { renderComponent(); }); - it('renders Stop Analysis button', () => { + it('renders Stop Analysis button', async () => { const btn = screen.getByRole('button', { name: 'Stop Analysis' }); expect(btn).toBeInTheDocument(); - userEvent.click(btn); + await userEvent.click(btn); expect(cancelMock).toHaveBeenCalled(); }); @@ -62,6 +63,10 @@ describe('Metrics', () => { expect(progressbar).toBeInTheDocument(); expect(progressbar).toHaveStyleRule('width', '0%'); }); + + it('calculate Timer ', () => { + expect(screen.getByText('Passed since start')).toBeInTheDocument(); + }); }); describe('when analysis is completed', () => { diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Statistics/__test__/Statistics.spec.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Statistics/__test__/Statistics.spec.tsx index d69c68b8edb..ef034b4d83a 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Statistics/__test__/Statistics.spec.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Statistics/__test__/Statistics.spec.tsx @@ -1,9 +1,10 @@ import React from 'react'; -import { screen } from '@testing-library/react'; +import { screen, waitFor } from '@testing-library/react'; import { render, WithRoute } from 'lib/testHelpers'; import Statistics from 'components/Topics/Topic/Statistics/Statistics'; import { clusterTopicStatisticsPath } from 'lib/paths'; -import { useTopicAnalysis } from 'lib/hooks/api/topics'; +import { useTopicAnalysis, useAnalyzeTopic } from 'lib/hooks/api/topics'; +import userEvent from '@testing-library/user-event'; const clusterName = 'local'; const topicName = 'topic'; @@ -11,6 +12,7 @@ const topicName = 'topic'; jest.mock('lib/hooks/api/topics', () => ({ ...jest.requireActual('lib/hooks/api/topics'), useTopicAnalysis: jest.fn(), + useAnalyzeTopic: jest.fn(), })); describe('Statistics', () => { @@ -23,13 +25,24 @@ describe('Statistics', () => { { initialEntries: [path] } ); }; - - it('renders Metricks component', () => { + const startMock = jest.fn(); + it('renders Metricks component', async () => { (useTopicAnalysis as jest.Mock).mockImplementation(() => ({ data: { result: 1 }, })); renderComponent(); - expect(screen.getByText('Restart Analysis')).toBeInTheDocument(); + await expect(screen.getByText('Restart Analysis')).toBeInTheDocument(); + expect(screen.queryByRole('progressbar')).not.toBeInTheDocument(); + }); + it('renders Start Analysis button', async () => { + (useAnalyzeTopic as jest.Mock).mockImplementation(() => ({ + mutateAsync: startMock, + })); + renderComponent(); + const btn = screen.getByRole('button', { name: 'Start Analysis' }); + expect(btn).toBeInTheDocument(); + await waitFor(() => userEvent.click(btn)); + expect(startMock).toHaveBeenCalled(); }); }); diff --git a/kafka-ui-react-app/src/lib/__test__/dateTimeHelpers.spec.ts b/kafka-ui-react-app/src/lib/__test__/dateTimeHelpers.spec.ts new file mode 100644 index 00000000000..c34371452b6 --- /dev/null +++ b/kafka-ui-react-app/src/lib/__test__/dateTimeHelpers.spec.ts @@ -0,0 +1,51 @@ +import { + passedTime, + calculateTimer, + formatMilliseconds, +} from 'lib/dateTimeHelpers'; + +const startedAt = 1664891890889; + +describe('format Milliseconds', () => { + it('hours > 0', () => { + const result = formatMilliseconds(10000000); + + expect(result).toEqual('2h 46m'); + }); + it('minutes > 0', () => { + const result = formatMilliseconds(1000000); + + expect(result).toEqual('16m 40s'); + }); + + it('seconds > 0', () => { + const result = formatMilliseconds(10000); + + expect(result).toEqual('10s'); + }); + + it('milliseconds > 0', () => { + const result = formatMilliseconds(100); + + expect(result).toEqual('100ms' || '0ms'); + expect(formatMilliseconds()).toEqual('0ms'); + }); +}); + +describe('calculate timer', () => { + it('time value < 10', () => { + expect(passedTime(5)).toBeTruthy(); + }); + + it('time value > 9', () => { + expect(passedTime(10)).toBeTruthy(); + }); + + it('run calculate time', () => { + expect(calculateTimer(startedAt)); + }); + + it('return when startedAt > new Date()', () => { + expect(calculateTimer(1664891890889199)).toBe('00:00'); + }); +}); diff --git a/kafka-ui-react-app/src/lib/dateTimeHelpers.ts b/kafka-ui-react-app/src/lib/dateTimeHelpers.ts index 57a8834ad3a..5c72af7bc47 100644 --- a/kafka-ui-react-app/src/lib/dateTimeHelpers.ts +++ b/kafka-ui-react-app/src/lib/dateTimeHelpers.ts @@ -32,3 +32,14 @@ export const formatMilliseconds = (input = 0) => { return `${milliseconds}ms`; }; + +export const passedTime = (value: number) => (value < 10 ? `0${value}` : value); + +export const calculateTimer = (startedAt: number) => { + const now = new Date().getTime(); + const newDate = now - startedAt; + const minutes = dayjs(newDate).minute(); + const second = dayjs(newDate).second(); + + return newDate > 0 ? `${passedTime(minutes)}:${passedTime(second)}` : '00:00'; +};
null
train
val
2022-11-04T20:11:08
"2022-09-30T09:18:27Z"
Haarolean
train
provectus/kafka-ui/2650_2692
provectus/kafka-ui
provectus/kafka-ui/2650
provectus/kafka-ui/2692
[ "keyword_pr_to_issue" ]
55c8d6ac2d4bdfb9721c79d4894732c86e0380d9
8731de1520705c29bfacbdf78c362912c6945132
[]
[]
"2022-10-01T12:10:09Z"
[ "type/bug", "good first issue", "scope/frontend", "status/accepted", "status/confirmed", "hacktoberfest" ]
Brokers: Broker config "Edit" button is either partially or not shown at all on narrow window size.
**Set up** Version: [eb06235](https://github.com/provectus/kafka-ui/commit/eb06235) **Steps to Reproduce** 1. Open the "Brokers" page and click on one of the brokers. 2. Open the Configs tab. **Actual behavior** - If the window width is narrower than 1284px, then "Edit" button is not fitting in the window, and not visible at all at 1215px or less. - Source tab is not visible at all in that case. - No horizontal scroll is available on the page. **Expected behavior** - "Key" and "Value" columns have a lot of free space and could be narrowed down to value length or value broken down into several lines, if they are too wide. - Horizontal scroll should be added, if the suggestion above is not possible. **Screenshots** ![image](https://user-images.githubusercontent.com/112083452/192478027-fc5ffb54-a9a3-4a1f-b9da-2ec5149fabe7.png) **Additional context** Might be affected by "Broker configs: allow resetting values to nulls [#2582](https://github.com/provectus/kafka-ui/issues/2582)"
[ "kafka-ui-react-app/src/components/App.styled.ts", "kafka-ui-react-app/src/components/Brokers/Broker/Configs/Configs.styled.ts", "kafka-ui-react-app/src/components/Brokers/Broker/Configs/InputCell.tsx", "kafka-ui-react-app/src/components/Topics/Topic/MessagesV2/Messages.styled.ts", "kafka-ui-react-app/src/components/common/NewTable/Table.styled.ts", "kafka-ui-react-app/src/components/common/NewTable/Table.tsx", "kafka-ui-react-app/src/theme/theme.ts" ]
[ "kafka-ui-react-app/src/components/App.styled.ts", "kafka-ui-react-app/src/components/Brokers/Broker/Configs/Configs.styled.ts", "kafka-ui-react-app/src/components/Brokers/Broker/Configs/InputCell.tsx", "kafka-ui-react-app/src/components/Topics/Topic/MessagesV2/Messages.styled.ts", "kafka-ui-react-app/src/components/common/NewTable/Table.styled.ts", "kafka-ui-react-app/src/components/common/NewTable/Table.tsx", "kafka-ui-react-app/src/theme/theme.ts" ]
[]
diff --git a/kafka-ui-react-app/src/components/App.styled.ts b/kafka-ui-react-app/src/components/App.styled.ts index 71353f08762..90eade047d1 100644 --- a/kafka-ui-react-app/src/components/App.styled.ts +++ b/kafka-ui-react-app/src/components/App.styled.ts @@ -20,9 +20,10 @@ export const Container = styled.main( position: relative; padding-bottom: 30px; z-index: 20; - + max-width: calc(100vw - ${theme.layout.navBarWidth}); @media screen and (max-width: 1023px) { margin-left: initial; + max-width: 100vw; } ` ); diff --git a/kafka-ui-react-app/src/components/Brokers/Broker/Configs/Configs.styled.ts b/kafka-ui-react-app/src/components/Brokers/Broker/Configs/Configs.styled.ts index a49bcf5d7d9..03030b12498 100644 --- a/kafka-ui-react-app/src/components/Brokers/Broker/Configs/Configs.styled.ts +++ b/kafka-ui-react-app/src/components/Brokers/Broker/Configs/Configs.styled.ts @@ -12,7 +12,7 @@ export const Value = styled.span` line-height: 24px; margin-right: 10px; text-overflow: ellipsis; - width: 600px; + max-width: 400px; overflow: hidden; white-space: nowrap; `; diff --git a/kafka-ui-react-app/src/components/Brokers/Broker/Configs/InputCell.tsx b/kafka-ui-react-app/src/components/Brokers/Broker/Configs/InputCell.tsx index ceeff2bd59e..422abbbac99 100644 --- a/kafka-ui-react-app/src/components/Brokers/Broker/Configs/InputCell.tsx +++ b/kafka-ui-react-app/src/components/Brokers/Broker/Configs/InputCell.tsx @@ -70,7 +70,7 @@ const InputCell: React.FC<InputCellProps> = ({ row, getValue, onUpdate }) => { : { fontWeight: 400 } } > - <S.Value>{initialValue}</S.Value> + <S.Value title={initialValue}>{initialValue}</S.Value> <Button buttonType="primary" buttonSize="S" diff --git a/kafka-ui-react-app/src/components/Topics/Topic/MessagesV2/Messages.styled.ts b/kafka-ui-react-app/src/components/Topics/Topic/MessagesV2/Messages.styled.ts index da464fe808f..c360f706d37 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/MessagesV2/Messages.styled.ts +++ b/kafka-ui-react-app/src/components/Topics/Topic/MessagesV2/Messages.styled.ts @@ -14,7 +14,7 @@ export const Wrapper = styled.div( export const Sidebar = styled.div( ({ theme }) => css` - width: 300px; + width: ${theme.layout.filtersSidebarWidth}; position: sticky; top: ${theme.layout.navBarHeight}; align-self: start; @@ -27,7 +27,9 @@ export const SidebarContent = styled.div` export const TableWrapper = styled.div( ({ theme }) => css` - width: 100%; + width: calc( + 100vw - ${theme.layout.navBarWidth} - ${theme.layout.filtersSidebarWidth} + ); border-left: 1px solid ${theme.layout.stuffBorderColor}; ` ); diff --git a/kafka-ui-react-app/src/components/common/NewTable/Table.styled.ts b/kafka-ui-react-app/src/components/common/NewTable/Table.styled.ts index 483e4e32160..d66bb45091c 100644 --- a/kafka-ui-react-app/src/components/common/NewTable/Table.styled.ts +++ b/kafka-ui-react-app/src/components/common/NewTable/Table.styled.ts @@ -211,3 +211,7 @@ export const Ellipsis = styled.div` text-overflow: ellipsis; display: block; `; + +export const TableWrapper = styled.div` + overflow-x: auto; +`; diff --git a/kafka-ui-react-app/src/components/common/NewTable/Table.tsx b/kafka-ui-react-app/src/components/common/NewTable/Table.tsx index 2d1646aa99e..7e3b571c375 100644 --- a/kafka-ui-react-app/src/components/common/NewTable/Table.tsx +++ b/kafka-ui-react-app/src/components/common/NewTable/Table.tsx @@ -198,95 +198,97 @@ const Table: React.FC<TableProps<any>> = ({ /> </S.TableActionsBar> )} - <S.Table> - <thead> - {table.getHeaderGroups().map((headerGroup) => ( - <tr key={headerGroup.id}> - {!!enableRowSelection && ( - <S.Th key={`${headerGroup.id}-select`}> - {flexRender( - SelectRowHeader, - headerGroup.headers[0].getContext() - )} - </S.Th> - )} - {table.getCanSomeRowsExpand() && ( - <S.Th expander key={`${headerGroup.id}-expander`} /> - )} - {headerGroup.headers.map((header) => ( - <S.Th - key={header.id} - colSpan={header.colSpan} - sortable={header.column.getCanSort()} - sortOrder={header.column.getIsSorted()} - onClick={header.column.getToggleSortingHandler()} - > - <div> - {flexRender( - header.column.columnDef.header, - header.getContext() - )} - </div> - </S.Th> - ))} - </tr> - ))} - </thead> - <tbody> - {table.getRowModel().rows.map((row) => ( - <React.Fragment key={row.id}> - <S.Row - expanded={row.getIsExpanded()} - onClick={handleRowClick(row)} - clickable={ - !enableRowSelection && - (row.getCanExpand() || onRowClick !== undefined) - } - > + <S.TableWrapper> + <S.Table> + <thead> + {table.getHeaderGroups().map((headerGroup) => ( + <tr key={headerGroup.id}> {!!enableRowSelection && ( - <td key={`${row.id}-select`} style={{ width: '1px' }}> + <S.Th key={`${headerGroup.id}-select`}> {flexRender( - SelectRowCell, - row.getVisibleCells()[0].getContext() + SelectRowHeader, + headerGroup.headers[0].getContext() )} - </td> + </S.Th> )} {table.getCanSomeRowsExpand() && ( - <td key={`${row.id}-expander`} style={{ width: '1px' }}> - {flexRender( - ExpanderCell, - row.getVisibleCells()[0].getContext() - )} - </td> + <S.Th expander key={`${headerGroup.id}-expander`} /> )} - {row - .getVisibleCells() - .map(({ id, getContext, column: { columnDef } }) => ( - <td key={id} style={columnDef.meta}> - {flexRender(columnDef.cell, getContext())} + {headerGroup.headers.map((header) => ( + <S.Th + key={header.id} + colSpan={header.colSpan} + sortable={header.column.getCanSort()} + sortOrder={header.column.getIsSorted()} + onClick={header.column.getToggleSortingHandler()} + > + <div> + {flexRender( + header.column.columnDef.header, + header.getContext() + )} + </div> + </S.Th> + ))} + </tr> + ))} + </thead> + <tbody> + {table.getRowModel().rows.map((row) => ( + <React.Fragment key={row.id}> + <S.Row + expanded={row.getIsExpanded()} + onClick={handleRowClick(row)} + clickable={ + !enableRowSelection && + (row.getCanExpand() || onRowClick !== undefined) + } + > + {!!enableRowSelection && ( + <td key={`${row.id}-select`} style={{ width: '1px' }}> + {flexRender( + SelectRowCell, + row.getVisibleCells()[0].getContext() + )} </td> - ))} - </S.Row> - {row.getIsExpanded() && SubComponent && ( - <S.Row expanded> - <td colSpan={row.getVisibleCells().length + 2}> - <S.ExpandedRowInfo> - <SubComponent row={row} /> - </S.ExpandedRowInfo> - </td> + )} + {table.getCanSomeRowsExpand() && ( + <td key={`${row.id}-expander`} style={{ width: '1px' }}> + {flexRender( + ExpanderCell, + row.getVisibleCells()[0].getContext() + )} + </td> + )} + {row + .getVisibleCells() + .map(({ id, getContext, column: { columnDef } }) => ( + <td key={id} style={columnDef.meta}> + {flexRender(columnDef.cell, getContext())} + </td> + ))} </S.Row> - )} - </React.Fragment> - ))} - {table.getRowModel().rows.length === 0 && ( - <S.Row> - <S.EmptyTableMessageCell colSpan={100}> - {emptyMessage || 'No rows found'} - </S.EmptyTableMessageCell> - </S.Row> - )} - </tbody> - </S.Table> + {row.getIsExpanded() && SubComponent && ( + <S.Row expanded> + <td colSpan={row.getVisibleCells().length + 2}> + <S.ExpandedRowInfo> + <SubComponent row={row} /> + </S.ExpandedRowInfo> + </td> + </S.Row> + )} + </React.Fragment> + ))} + {table.getRowModel().rows.length === 0 && ( + <S.Row> + <S.EmptyTableMessageCell colSpan={100}> + {emptyMessage || 'No rows found'} + </S.EmptyTableMessageCell> + </S.Row> + )} + </tbody> + </S.Table> + </S.TableWrapper> {table.getPageCount() > 1 && ( <S.Pagination> <S.Pages> diff --git a/kafka-ui-react-app/src/theme/theme.ts b/kafka-ui-react-app/src/theme/theme.ts index 19c082c2840..5a2b806a01f 100644 --- a/kafka-ui-react-app/src/theme/theme.ts +++ b/kafka-ui-react-app/src/theme/theme.ts @@ -90,6 +90,7 @@ const theme = { navBarWidth: '201px', navBarHeight: '53px', rightSidebarWidth: '70vw', + filtersSidebarWidth: '300px', stuffColor: Colors.neutral[5], stuffBorderColor: Colors.neutral[10],
null
train
val
2022-10-02T09:34:13
"2022-09-27T08:48:09Z"
BulatKha
train
provectus/kafka-ui/2682_2694
provectus/kafka-ui
provectus/kafka-ui/2682
provectus/kafka-ui/2694
[ "timestamp(timedelta=1.0, similarity=0.9584831007797872)", "connected" ]
c9613448242ff85c5f83361ec0c381ef39c37e58
0939d6140f5540bd421f19d61d862d363cff7456
[]
[ "I don't like the approach when we create toipc link component. Let's try to create universal component where we will send `to` as a property ", "Not sure why we need to remove this test case", "let's call it LinkCell and update all the existing tables with same link components", "All the link colours must be the same\r\n```suggestion\r\n <NavLink to={to} title={value} onClick={handleClick}>\r\n```", "```suggestion\r\n \r\n```" ]
"2022-10-02T00:07:02Z"
[ "good first issue", "scope/frontend", "type/refactoring" ]
Migrate Dashboard table to the new version on table components
### Describe the solution you'd like Migrate `src/components/Dashboard/ClustersWidget/ClustersWidget.tsx` to new version of tables `src/components/common/NewTable/Table.tsx` ### Additional context Go to Dasboard
[ "kafka-ui-react-app/src/components/Brokers/BrokersList/BrokersList.tsx", "kafka-ui-react-app/src/components/ConsumerGroups/List/List.tsx", "kafka-ui-react-app/src/components/Dashboard/ClustersWidget/ClustersWidget.tsx", "kafka-ui-react-app/src/components/Dashboard/ClustersWidget/__test__/ClustersWidget.spec.tsx", "kafka-ui-react-app/src/components/Schemas/List/List.tsx", "kafka-ui-react-app/src/components/common/NewTable/LinkCell.tsx", "kafka-ui-react-app/src/components/common/NewTable/__test__/Table.spec.tsx" ]
[ "kafka-ui-react-app/src/components/Brokers/BrokersList/BrokersList.tsx", "kafka-ui-react-app/src/components/ConsumerGroups/List/List.tsx", "kafka-ui-react-app/src/components/Dashboard/ClustersWidget/ClusterName.tsx", "kafka-ui-react-app/src/components/Dashboard/ClustersWidget/ClustersWidget.tsx", "kafka-ui-react-app/src/components/Dashboard/ClustersWidget/__test__/ClustersWidget.spec.tsx", "kafka-ui-react-app/src/components/Schemas/List/List.tsx", "kafka-ui-react-app/src/components/common/NewTable/LinkCell.tsx", "kafka-ui-react-app/src/components/common/NewTable/__test__/Table.spec.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Brokers/BrokersList/BrokersList.tsx b/kafka-ui-react-app/src/components/Brokers/BrokersList/BrokersList.tsx index 9f63be31c29..54bdcef5ff8 100644 --- a/kafka-ui-react-app/src/components/Brokers/BrokersList/BrokersList.tsx +++ b/kafka-ui-react-app/src/components/Brokers/BrokersList/BrokersList.tsx @@ -42,10 +42,19 @@ const BrokersList: React.FC = () => { }; }); }, [diskUsage, brokers]); - const columns = React.useMemo<ColumnDef<typeof rows>[]>( () => [ - { header: 'Broker ID', accessorKey: 'brokerId', cell: LinkCell }, + { + header: 'Broker ID', + accessorKey: 'brokerId', + // eslint-disable-next-line react/no-unstable-nested-components + cell: ({ getValue }) => ( + <LinkCell + value={`${getValue<string | number>()}`} + to={encodeURIComponent(`${getValue<string | number>()}`)} + /> + ), + }, { header: 'Segment Size', accessorKey: 'size', cell: SizeCell }, { header: 'Segment Count', accessorKey: 'count' }, { header: 'Port', accessorKey: 'port' }, diff --git a/kafka-ui-react-app/src/components/ConsumerGroups/List/List.tsx b/kafka-ui-react-app/src/components/ConsumerGroups/List/List.tsx index 89be57d6392..c7df01b5e95 100644 --- a/kafka-ui-react-app/src/components/ConsumerGroups/List/List.tsx +++ b/kafka-ui-react-app/src/components/ConsumerGroups/List/List.tsx @@ -49,7 +49,13 @@ const List: React.FC<Props> = ({ consumerGroups, totalPages }) => { id: ConsumerGroupOrdering.NAME, header: 'Group ID', accessorKey: 'groupId', - cell: LinkCell, + // eslint-disable-next-line react/no-unstable-nested-components + cell: ({ getValue }) => ( + <LinkCell + value={`${getValue<string | number>()}`} + to={encodeURIComponent(`${getValue<string | number>()}`)} + /> + ), }, { id: ConsumerGroupOrdering.MEMBERS, diff --git a/kafka-ui-react-app/src/components/Dashboard/ClustersWidget/ClusterName.tsx b/kafka-ui-react-app/src/components/Dashboard/ClustersWidget/ClusterName.tsx new file mode 100644 index 00000000000..52702c348a7 --- /dev/null +++ b/kafka-ui-react-app/src/components/Dashboard/ClustersWidget/ClusterName.tsx @@ -0,0 +1,15 @@ +import React from 'react'; +import { CellContext } from '@tanstack/react-table'; +import { Tag } from 'components/common/Tag/Tag.styled'; + +// eslint-disable-next-line @typescript-eslint/no-explicit-any +const ClusterName: React.FC<CellContext<any, unknown>> = ({ row }) => { + return ( + <> + {row.original.readOnly && <Tag color="blue">readonly</Tag>} + {row.original.name} + </> + ); +}; + +export default ClusterName; diff --git a/kafka-ui-react-app/src/components/Dashboard/ClustersWidget/ClustersWidget.tsx b/kafka-ui-react-app/src/components/Dashboard/ClustersWidget/ClustersWidget.tsx index 7e303c98763..0327a010012 100644 --- a/kafka-ui-react-app/src/components/Dashboard/ClustersWidget/ClustersWidget.tsx +++ b/kafka-ui-react-app/src/components/Dashboard/ClustersWidget/ClustersWidget.tsx @@ -1,16 +1,14 @@ import React from 'react'; import * as Metrics from 'components/common/Metrics'; import { Tag } from 'components/common/Tag/Tag.styled'; -import { Table } from 'components/common/table/Table/Table.styled'; -import TableHeaderCell from 'components/common/table/TableHeaderCell/TableHeaderCell'; -import BytesFormatted from 'components/common/BytesFormatted/BytesFormatted'; -import { NavLink } from 'react-router-dom'; -import { clusterTopicsPath } from 'lib/paths'; import Switch from 'components/common/Switch/Switch'; import { useClusters } from 'lib/hooks/api/clusters'; -import { ServerStatus } from 'generated-sources'; +import { Cluster, ServerStatus } from 'generated-sources'; +import { ColumnDef } from '@tanstack/react-table'; +import Table, { SizeCell } from 'components/common/NewTable'; import * as S from './ClustersWidget.styled'; +import ClusterName from './ClusterName'; const ClustersWidget: React.FC = () => { const { data } = useClusters(); @@ -28,6 +26,19 @@ const ClustersWidget: React.FC = () => { }; }, [data, showOfflineOnly]); + const columns = React.useMemo<ColumnDef<Cluster>[]>( + () => [ + { header: 'Cluster name', accessorKey: 'name', cell: ClusterName }, + { header: 'Version', accessorKey: 'version' }, + { header: 'Brokers count', accessorKey: 'brokerCount' }, + { header: 'Partitions', accessorKey: 'onlinePartitionCount' }, + { header: 'Topics', accessorKey: 'topicCount' }, + { header: 'Production', accessorKey: 'bytesInPerSec', cell: SizeCell }, + { header: 'Consumption', accessorKey: 'bytesOutPerSec', cell: SizeCell }, + ], + [] + ); + const handleSwitch = () => setShowOfflineOnly(!showOfflineOnly); return ( <> @@ -51,45 +62,12 @@ const ClustersWidget: React.FC = () => { /> <label>Only offline clusters</label> </S.SwitchWrapper> - <Table isFullwidth> - <thead> - <tr> - <TableHeaderCell title="Cluster name" /> - <TableHeaderCell title="Version" /> - <TableHeaderCell title="Brokers count" /> - <TableHeaderCell title="Partitions" /> - <TableHeaderCell title="Topics" /> - <TableHeaderCell title="Production" /> - <TableHeaderCell title="Consumption" /> - </tr> - </thead> - <tbody> - {config.list.map((cluster) => ( - <tr key={cluster.name}> - <S.TableCell maxWidth="99px" width="350"> - {cluster.readOnly && <Tag color="blue">readonly</Tag>}{' '} - {cluster.name} - </S.TableCell> - <S.TableCell maxWidth="99px">{cluster.version}</S.TableCell> - <S.TableCell maxWidth="99px">{cluster.brokerCount}</S.TableCell> - <S.TableCell maxWidth="78px"> - {cluster.onlinePartitionCount} - </S.TableCell> - <S.TableCell maxWidth="60px"> - <NavLink to={clusterTopicsPath(cluster.name)}> - {cluster.topicCount} - </NavLink> - </S.TableCell> - <S.TableCell maxWidth="85px"> - <BytesFormatted value={cluster.bytesInPerSec} /> - </S.TableCell> - <S.TableCell maxWidth="85px"> - <BytesFormatted value={cluster.bytesOutPerSec} /> - </S.TableCell> - </tr> - ))} - </tbody> - </Table> + <Table + columns={columns} + data={config?.list} + enableSorting + emptyMessage="Disk usage data not available" + /> </> ); }; diff --git a/kafka-ui-react-app/src/components/Dashboard/ClustersWidget/__test__/ClustersWidget.spec.tsx b/kafka-ui-react-app/src/components/Dashboard/ClustersWidget/__test__/ClustersWidget.spec.tsx index c5fa81342c6..dfdcb341796 100644 --- a/kafka-ui-react-app/src/components/Dashboard/ClustersWidget/__test__/ClustersWidget.spec.tsx +++ b/kafka-ui-react-app/src/components/Dashboard/ClustersWidget/__test__/ClustersWidget.spec.tsx @@ -36,8 +36,7 @@ describe('ClustersWidget', () => { }); it('render clusterWidget cells', () => { - const cells = screen.getAllByRole('cells'); + const cells = screen.getAllByRole('cell'); expect(cells.length).toBe(14); - expect(cells[0]).toHaveStyle('max-width: 99px'); }); }); diff --git a/kafka-ui-react-app/src/components/Schemas/List/List.tsx b/kafka-ui-react-app/src/components/Schemas/List/List.tsx index 31c280b92ad..8999a81d7b4 100644 --- a/kafka-ui-react-app/src/components/Schemas/List/List.tsx +++ b/kafka-ui-react-app/src/components/Schemas/List/List.tsx @@ -54,7 +54,17 @@ const List: React.FC = () => { const columns = React.useMemo<ColumnDef<SchemaSubject>[]>( () => [ - { header: 'Subject', accessorKey: 'subject', cell: LinkCell }, + { + header: 'Subject', + accessorKey: 'subject', + // eslint-disable-next-line react/no-unstable-nested-components + cell: ({ getValue }) => ( + <LinkCell + value={`${getValue<string | number>()}`} + to={encodeURIComponent(`${getValue<string | number>()}`)} + /> + ), + }, { header: 'Id', accessorKey: 'id' }, { header: 'Type', accessorKey: 'schemaType' }, { header: 'Version', accessorKey: 'version' }, diff --git a/kafka-ui-react-app/src/components/common/NewTable/LinkCell.tsx b/kafka-ui-react-app/src/components/common/NewTable/LinkCell.tsx index b6ca656d1d1..e400fa0b9c3 100644 --- a/kafka-ui-react-app/src/components/common/NewTable/LinkCell.tsx +++ b/kafka-ui-react-app/src/components/common/NewTable/LinkCell.tsx @@ -1,13 +1,11 @@ import React from 'react'; -import { CellContext } from '@tanstack/react-table'; import { NavLink } from 'react-router-dom'; // eslint-disable-next-line @typescript-eslint/no-explicit-any -const LinkCell: React.FC<CellContext<any, unknown>> = ({ getValue }) => { - const value = `${getValue<string | number>()}`; +const LinkCell = ({ value, to = '' }: any) => { const handleClick: React.MouseEventHandler = (e) => e.stopPropagation(); return ( - <NavLink to={encodeURIComponent(value)} title={value} onClick={handleClick}> + <NavLink to={to} title={value} onClick={handleClick}> {value} </NavLink> ); diff --git a/kafka-ui-react-app/src/components/common/NewTable/__test__/Table.spec.tsx b/kafka-ui-react-app/src/components/common/NewTable/__test__/Table.spec.tsx index 412fafd484c..480f6e6db15 100644 --- a/kafka-ui-react-app/src/components/common/NewTable/__test__/Table.spec.tsx +++ b/kafka-ui-react-app/src/components/common/NewTable/__test__/Table.spec.tsx @@ -63,7 +63,12 @@ const columns: ColumnDef<Datum>[] = [ { header: 'Text', accessorKey: 'text', - cell: LinkCell, + cell: ({ getValue }) => ( + <LinkCell + value={`${getValue<string | number>()}`} + to={encodeURIComponent(`${getValue<string | number>()}`)} + /> + ), }, { header: 'Size',
null
train
val
2022-10-04T14:31:07
"2022-09-30T10:45:53Z"
workshur
train
provectus/kafka-ui/2642_2701
provectus/kafka-ui
provectus/kafka-ui/2642
provectus/kafka-ui/2701
[ "timestamp(timedelta=1.0, similarity=0.9396692525091919)", "connected" ]
6dff8f105ec55d04bf5018232dcb25c6b83d3e8b
67eea972f77c1e133e9e7360740189a0fa3373f2
[]
[ "why do we need this annotation?", "$x(\"//h1[text()='Dashboard']\") is it Dashboard title?\nwe need to wait NaviSideBar element in this method, I meant to get Dashboard menu", "please don't place enums between steps\nplease set waitUntilScreenReady() at the top of methods\nenums can be moved to the footer", "because of it's the first waiter that we met after opening URL we need to increase Duration in aviSideBar.waitUntilScreenReady(). I believe we can set 30 sec to try from start", "can we use static import NaviSideBar.SideMenuOptions.*?", "kafkaConnectList.waitUntilScreenReady() missed", "let's simplify this method. today we have only one cluster so we can overload method only with SideMenuOption as argument and use it in test. please make such method with hardcoded clusterName", "use single number, pls: SideMenuOptions -> SideMenuOption", "schemaRegistryList.waitUntilScreenReady() missed", "schemaRegistryList.waitUntilScreenReady() missed", "schemaRegistryList.waitUntilScreenReady() missed", "schemaRegistryList.waitUntilScreenReady() missed", "schemaRegistryList.waitUntilScreenReady() missed", "schemaRegistryList.waitUntilScreenReady() missed", "topicsList.waitUntilScreenReady() missed", "Added", "Added", "Added", "Added", "Added", "Added", "Added", "Fixed", "Done.", "Added", "Fixed", "Added", "Fixed", "Fixed", "Fixed", "where??", "it's bad practice to use some strange spaces in locators\nbetter use 'contains' in this case", "extra space, remove pls", "partially, take a look again pls", "kafkaConnectList.waitUntilScreenReady() missed", "kafkaConnectList.waitUntilScreenReady() missed", "schemaRegistryList.waitUntilScreenReady() missed", "insert pls schemaDetails object to understand which screen are we waiting for", "schemaRegistryList.waitUntilScreenReady() missed", "schemaRegistryList.waitUntilScreenReady() missed", "schemaRegistryList.waitUntilScreenReady() missed", "schemaRegistryList.waitUntilScreenReady() missed", "schemaRegistryList.waitUntilScreenReady() missed", "topicsList.waitUntilScreenReady() missed", "Fixed", "Have done!", "Fixed", "Fixed", "Fixed", "Added", "Added", "Added", "Added", "Added", "Added", "Added", "Added", "where??", "after openEditSchema() we should wait schemaCreateForm", "Have done!", "Added" ]
"2022-10-04T14:14:27Z"
[ "type/enhancement", "scope/QA", "status/accepted" ]
[e2e] create navigation thru the app using web driver
- create NaviSideBar class - implement all navigations
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/MainPage.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaCreateForm.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaRegistryList.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/ProduceMessagePanel.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/settings/Source.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/NaviSideBar.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaCreateForm.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaRegistryList.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/ProduceMessagePanel.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/settings/Source.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/SmokeTests.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/BaseTest.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Setup.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/ConnectorsTests.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/SchemasTests.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/MainPage.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/NaviSideBar.java similarity index 50% rename from kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/MainPage.java rename to kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/NaviSideBar.java index ba93f0535c7..30796617122 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/MainPage.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/NaviSideBar.java @@ -1,58 +1,55 @@ package com.provectus.kafka.ui.pages; import com.codeborne.selenide.Condition; -import com.codeborne.selenide.Selenide; import com.codeborne.selenide.SelenideElement; -import com.provectus.kafka.ui.utilities.WaitUtils; -import com.provectus.kafka.ui.settings.Source; import io.qameta.allure.Step; -import lombok.experimental.ExtensionMethod; -import org.openqa.selenium.By; -import static com.codeborne.selenide.Selenide.$; +import java.time.Duration; + import static com.codeborne.selenide.Selenide.$x; +import static com.provectus.kafka.ui.settings.Source.CLUSTER_NAME; -@ExtensionMethod({WaitUtils.class}) -public class MainPage { +public class NaviSideBar { - private static final String path = "/"; + @Step + public NaviSideBar waitUntilScreenReady() { + $x("//*[contains(text(),'Loading')]").shouldBe(Condition.disappear, Duration.ofSeconds(30)); + $x("//a[@title='Dashboard']").shouldBe(Condition.visible, Duration.ofSeconds(30)); + return this; + } @Step - public MainPage goTo() { - Selenide.open(Source.BASE_WEB_URL + path); + public NaviSideBar openSideMenu(String clusterName, SideMenuOption option) { + SelenideElement clusterElement = $x(String.format("//aside/ul/li[contains(.,'%s')]", clusterName)).shouldBe(Condition.visible); + if (clusterElement.parent().$$x(".//ul").size() == 0) { + clusterElement.click(); + } + clusterElement + .parent() + .$x(String.format(".//ul/li[contains(.,'%s')]", option.value)) + .click(); return this; } @Step - public MainPage waitUntilScreenReady() { - $(By.xpath("//*[contains(text(),'Loading')]")).shouldBe(Condition.disappear); - $("input[name=switchRoundedDefault]").parent().$("span").shouldBe(Condition.visible); + public NaviSideBar openSideMenu(SideMenuOption option) { + openSideMenu(CLUSTER_NAME, option); return this; } - public enum SideMenuOptions { + + public enum SideMenuOption { BROKERS("Brokers"), TOPICS("Topics"), CONSUMERS("Consumers"), - SCHEMA_REGISTRY("Schema Registry"); + SCHEMA_REGISTRY("Schema Registry"), + KAFKA_CONNECT("Kafka Connect"), + KSQL_DB("KSQL DB"); final String value; - SideMenuOptions(String value) { + SideMenuOption(String value) { this.value = value; } } - - @Step - public MainPage goToSideMenu(String clusterName, SideMenuOptions option) { - SelenideElement clusterElement = $x(String.format("//aside/ul/li[contains(.,'%s')]", clusterName)).shouldBe(Condition.visible); - if (clusterElement.parent().$$x(".//ul").size() == 0) { - clusterElement.click(); - } - clusterElement - .parent() - .$x(String.format(".//ul/li[contains(.,'%s')]", option.value)) - .click(); - return this; - } -} +} \ No newline at end of file diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java index 70993072dfd..b5986c7919e 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java @@ -1,8 +1,6 @@ package com.provectus.kafka.ui.pages.connector; import com.codeborne.selenide.Condition; -import com.codeborne.selenide.Selenide; -import com.provectus.kafka.ui.settings.Source; import com.provectus.kafka.ui.utilities.WaitUtils; import io.qameta.allure.Step; import lombok.experimental.ExtensionMethod; @@ -16,14 +14,6 @@ @ExtensionMethod(WaitUtils.class) public class KafkaConnectList { - private static final String path = "/ui/clusters/%s/connectors"; - - @Step("Open URL to {cluster}") - public KafkaConnectList goTo(String cluster) { - Selenide.open(Source.BASE_WEB_URL + String.format(path, cluster)); - return this; - } - @Step public KafkaConnectList waitUntilScreenReady() { $(By.xpath("//h1[text()='Connectors']")).shouldBe(Condition.visible); diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaCreateForm.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaCreateForm.java index fec5db8b2e7..0f223ada78e 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaCreateForm.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaCreateForm.java @@ -1,5 +1,6 @@ package com.provectus.kafka.ui.pages.schema; +import com.codeborne.selenide.Condition; import com.codeborne.selenide.Selenide; import com.codeborne.selenide.SelenideElement; import com.provectus.kafka.ui.api.model.CompatibilityLevel; @@ -20,6 +21,12 @@ public class SchemaCreateForm { protected SelenideElement newSchemaTextArea = $("#newSchema [wrap]"); protected SelenideElement schemaTypeDropDown = $x("//ul[@name='schemaType']"); + @Step + public SchemaCreateForm waitUntilScreenReady() { + $x("//h1['Edit']").shouldBe(Condition.visible); + return this; + } + @Step public SchemaCreateForm setSubjectName(String name) { subjectName.setValue(name); diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaRegistryList.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaRegistryList.java index eb378be3db6..d37981d3044 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaRegistryList.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaRegistryList.java @@ -14,6 +14,13 @@ public class SchemaRegistryList { private final SelenideElement schemaButton = $(By.xpath("//*[contains(text(),'Create Schema')]")); + @Step + public SchemaRegistryList waitUntilScreenReady(){ + $x("//*[contains(text(),'Loading')]").shouldBe(Condition.disappear); + $x("//button[contains(text(),'Create Schema')]").shouldBe(Condition.visible); + return this; + } + @Step public SchemaCreateForm clickCreateSchema() { clickByJavaScript(schemaButton); diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/ProduceMessagePanel.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/ProduceMessagePanel.java index d6ecebcd7a7..62d619a5394 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/ProduceMessagePanel.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/ProduceMessagePanel.java @@ -43,6 +43,6 @@ public TopicDetails submitProduceMessage() { submitBtn.shouldBe(Condition.enabled).click(); submitBtn.shouldBe(Condition.disappear); refresh(); - return new TopicDetails().waitUntilScreenReady(); + return new TopicDetails(); } } diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java index e0400bea37b..123c3d76eec 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java @@ -1,9 +1,7 @@ package com.provectus.kafka.ui.pages.topic; import com.codeborne.selenide.Condition; -import com.codeborne.selenide.Selenide; import com.codeborne.selenide.SelenideElement; -import com.provectus.kafka.ui.settings.Source; import com.provectus.kafka.ui.utilities.WaitUtils; import io.qameta.allure.Step; import lombok.experimental.ExtensionMethod; @@ -15,15 +13,8 @@ @ExtensionMethod({WaitUtils.class}) public class TopicDetails { - private static final String URL_PATH = "/ui/clusters/%s/topics/%s"; protected SelenideElement dotMenuBtn = $$x("//button[@aria-label='Dropdown Toggle']").first(); - @Step - public TopicDetails goTo(String cluster, String topic) { - Selenide.open(Source.BASE_WEB_URL + String.format(URL_PATH, cluster, topic)); - return this; - } - @Step public TopicDetails waitUntilScreenReady() { $(By.linkText("Overview")).shouldBe(Condition.visible); @@ -70,7 +61,9 @@ public boolean isContentMessageVisible(String contentMessage) { } private enum DotMenuHeaderItems { - EDIT_SETTINGS("Edit settings"), CLEAR_MESSAGES("Clear messages"), REMOVE_TOPIC("Remove topic"); + EDIT_SETTINGS("Edit settings"), + CLEAR_MESSAGES("Clear messages"), + REMOVE_TOPIC("Remove topic"); private final String value; @@ -89,7 +82,10 @@ public String toString() { } public enum TopicMenu { - OVERVIEW("Overview"), MESSAGES("Messages"), CONSUMERS("Consumers"), SETTINGS("Settings"); + OVERVIEW("Overview"), + MESSAGES("Messages"), + CONSUMERS("Consumers"), + SETTINGS("Settings"); private final String value; diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java index 3412e0d5da8..0097bdb91a3 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java @@ -1,29 +1,19 @@ package com.provectus.kafka.ui.pages.topic; -import com.codeborne.selenide.CollectionCondition; import com.codeborne.selenide.Condition; -import com.codeborne.selenide.Selenide; -import com.provectus.kafka.ui.settings.Source; import com.provectus.kafka.ui.utilities.WaitUtils; import io.qameta.allure.Step; import lombok.experimental.ExtensionMethod; import org.openqa.selenium.By; -import static com.codeborne.selenide.Selenide.*; +import static com.codeborne.selenide.Selenide.$; +import static com.codeborne.selenide.Selenide.$x; import static com.provectus.kafka.ui.utilities.WebUtils.clickByJavaScript; import static com.provectus.kafka.ui.utilities.WebUtils.isVisible; @ExtensionMethod(WaitUtils.class) public class TopicsList { - private static final String path = "/ui/clusters/%s/all-topics"; - - @Step - public TopicsList goTo(String cluster) { - Selenide.open(Source.BASE_WEB_URL + String.format(path, cluster)); - return this; - } - @Step public TopicsList waitUntilScreenReady() { $(By.xpath("//*[contains(text(),'Loading')]")).shouldBe(Condition.disappear); @@ -48,14 +38,4 @@ public TopicDetails openTopic(String topicName) { $(By.linkText(topicName)).click(); return new TopicDetails(); } - - @Step - public TopicsList isTopicNotVisible(String topicName) { - $$x("//table/tbody/tr/td[2]") - .shouldBe(CollectionCondition.sizeGreaterThan(0)) - .find(Condition.exactText(topicName)) - .shouldBe(Condition.not(Condition.visible)); - return this; - } - } diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/settings/Source.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/settings/Source.java index f1c780214cb..0cbd7ae7c39 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/settings/Source.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/settings/Source.java @@ -4,4 +4,5 @@ public abstract class Source { public static String BASE_API_URL = System.getProperty("BASE_URL", "http://localhost:8080"); public static String BASE_WEB_URL = System.getProperty("BASE_DOCKER_URL", "http://host.testcontainers.internal:8080"); + public static final String CLUSTER_NAME = "local"; }
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/SmokeTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/SmokeTests.java index c9c560c3edc..e876d3a634b 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/SmokeTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/SmokeTests.java @@ -14,8 +14,6 @@ public class SmokeTests extends BaseTest { @CaseId(198) @DisplayName("main page should load") void mainPageLoads() { - mainPage.goTo() - .waitUntilScreenReady(); compareScreenshots("main"); } } diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/BaseTest.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/BaseTest.java index a864a36eb2c..a3086fd7151 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/BaseTest.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/BaseTest.java @@ -1,5 +1,7 @@ package com.provectus.kafka.ui.base; +import com.codeborne.selenide.Condition; +import com.codeborne.selenide.Selenide; import com.codeborne.selenide.WebDriverRunner; import com.provectus.kafka.ui.utilities.qaseIoUtils.DisplayNameGenerator; import com.provectus.kafka.ui.utilities.qaseIoUtils.TestCaseGenerator; @@ -22,9 +24,10 @@ import java.io.ByteArrayInputStream; import java.io.File; import java.io.IOException; +import java.time.Duration; -import static com.provectus.kafka.ui.base.Setup.clearReports; -import static com.provectus.kafka.ui.base.Setup.setup; +import static com.provectus.kafka.ui.base.Setup.*; +import static com.provectus.kafka.ui.settings.Source.BASE_WEB_URL; @Slf4j @DisplayNameGeneration(DisplayNameGenerator.class) @@ -32,8 +35,6 @@ public class BaseTest extends Facade { private static final String SELENIUM_IMAGE_NAME = "selenium/standalone-chrome:103.0"; private static final String SELENIARM_STANDALONE_CHROMIUM = "seleniarm/standalone-chromium:103.0"; - protected static final String CLUSTER_NAME = "local"; - private final Screenshooter screenshooter = new Screenshooter(); protected static BrowserWebDriverContainer<?> webDriverContainer = null; @@ -55,6 +56,8 @@ public void setWebDriver() { RemoteWebDriver remoteWebDriver = webDriverContainer.getWebDriver(); WebDriverRunner.setWebDriver(remoteWebDriver); remoteWebDriver.manage().window().setSize(new Dimension(1440, 1024)); + Selenide.open(BASE_WEB_URL); + naviSideBar.waitUntilScreenReady(); } @BeforeAll @@ -118,5 +121,6 @@ public void afterMethod() { Allure.addAttachment("Screenshot", new ByteArrayInputStream( ((TakesScreenshot) webDriverContainer.getWebDriver()).getScreenshotAs(OutputType.BYTES))); + browserClear(); } } diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java index 46e17e0b961..8d3aea55d4f 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java @@ -1,7 +1,7 @@ package com.provectus.kafka.ui.base; import com.provectus.kafka.ui.helpers.ApiHelper; -import com.provectus.kafka.ui.pages.MainPage; +import com.provectus.kafka.ui.pages.NaviSideBar; import com.provectus.kafka.ui.pages.connector.ConnectorCreateForm; import com.provectus.kafka.ui.pages.connector.ConnectorDetails; import com.provectus.kafka.ui.pages.connector.KafkaConnectList; @@ -14,7 +14,6 @@ import com.provectus.kafka.ui.pages.topic.TopicsList; public abstract class Facade { - protected MainPage mainPage = new MainPage(); protected ApiHelper apiHelper = new ApiHelper(); protected ConnectorCreateForm connectorCreateForm = new ConnectorCreateForm(); protected KafkaConnectList kafkaConnectList = new KafkaConnectList(); @@ -26,4 +25,5 @@ public abstract class Facade { protected TopicCreateEditForm topicCreateEditForm = new TopicCreateEditForm(); protected TopicsList topicsList = new TopicsList(); protected TopicDetails topicDetails = new TopicDetails(); + protected NaviSideBar naviSideBar = new NaviSideBar(); } diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Setup.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Setup.java index 954acb37ba7..2f64699eb26 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Setup.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Setup.java @@ -2,13 +2,19 @@ import com.codeborne.selenide.Configuration; import com.codeborne.selenide.logevents.SelenideLogger; +import io.qameta.allure.Step; import io.qameta.allure.selenide.AllureSelenide; import lombok.SneakyThrows; import lombok.extern.slf4j.Slf4j; +import org.junit.jupiter.api.AfterEach; +import org.junit.jupiter.api.BeforeEach; import java.io.File; import java.util.Arrays; +import static com.codeborne.selenide.Selenide.*; +import static com.provectus.kafka.ui.settings.Source.BASE_WEB_URL; + @Slf4j public abstract class Setup { @@ -38,4 +44,13 @@ public static void clearReports() { } } } + + @Step + public static void browserClear() { + log.debug("browserClear"); + clearBrowserLocalStorage(); + clearBrowserCookies(); + refresh(); + log.debug("=> DONE"); + } } diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/ConnectorsTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/ConnectorsTests.java index 4cc0c783e4b..bc69eeee6f0 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/ConnectorsTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/ConnectorsTests.java @@ -12,6 +12,8 @@ import java.util.ArrayList; import java.util.List; +import static com.provectus.kafka.ui.pages.NaviSideBar.SideMenuOption.KAFKA_CONNECT; +import static com.provectus.kafka.ui.settings.Source.CLUSTER_NAME; import static com.provectus.kafka.ui.utilities.FileUtils.getResourceAsString; @TestInstance(TestInstance.Lifecycle.PER_CLASS) @@ -60,14 +62,19 @@ public void createConnector() { Connector connectorForCreate = new Connector() .setName("sink_postgres_activities_e2e_checks") .setConfig(getResourceAsString("config_for_create_connector.json")); - kafkaConnectList.goTo(CLUSTER_NAME) + naviSideBar + .openSideMenu(KAFKA_CONNECT); + kafkaConnectList .waitUntilScreenReady() - .clickCreateConnectorButton() + .clickCreateConnectorButton(); + connectorCreateForm .waitUntilScreenReady() .setConnectorConfig(connectorForCreate.getName(), connectorForCreate.getConfig()); - kafkaConnectList.goTo(CLUSTER_NAME) + naviSideBar + .openSideMenu(KAFKA_CONNECT); + kafkaConnectList .waitUntilScreenReady(); - Assertions.assertTrue(kafkaConnectList.isConnectorVisible(connectorForCreate.getName()),"isConnectorVisible()"); + Assertions.assertTrue(kafkaConnectList.isConnectorVisible(connectorForCreate.getName()), "isConnectorVisible()"); CONNECTOR_LIST.add(connectorForCreate); } @@ -77,14 +84,20 @@ public void createConnector() { @CaseId(196) @Test public void updateConnector() { - kafkaConnectList.goTo(CLUSTER_NAME) + naviSideBar + .openSideMenu(KAFKA_CONNECT); + kafkaConnectList .waitUntilScreenReady() .openConnector(CONNECTOR_FOR_UPDATE.getName()); - connectorDetails.waitUntilScreenReady() + connectorDetails + .waitUntilScreenReady() .openConfigTab() .setConfig(CONNECTOR_FOR_UPDATE.getConfig()); - kafkaConnectList.goTo(CLUSTER_NAME); - Assertions.assertTrue(kafkaConnectList.isConnectorVisible(CONNECTOR_FOR_UPDATE.getName()),"isConnectorVisible()"); + naviSideBar + .openSideMenu(KAFKA_CONNECT); + kafkaConnectList + .waitUntilScreenReady(); + Assertions.assertTrue(kafkaConnectList.isConnectorVisible(CONNECTOR_FOR_UPDATE.getName()), "isConnectorVisible()"); } @DisplayName("should delete connector") @@ -93,7 +106,9 @@ public void updateConnector() { @CaseId(195) @Test public void deleteConnector() { - kafkaConnectList.goTo(CLUSTER_NAME) + naviSideBar + .openSideMenu(KAFKA_CONNECT); + kafkaConnectList .waitUntilScreenReady() .openConnector(CONNECTOR_FOR_DELETE.getName()); connectorDetails @@ -101,7 +116,10 @@ public void deleteConnector() { .openDotMenu() .clickDeleteButton() .clickConfirmButton(); - kafkaConnectList.goTo(CLUSTER_NAME); + naviSideBar + .openSideMenu(KAFKA_CONNECT); + kafkaConnectList + .waitUntilScreenReady(); Assertions.assertFalse(kafkaConnectList.isConnectorVisible(CONNECTOR_FOR_DELETE.getName()), "isConnectorVisible()"); CONNECTOR_LIST.remove(CONNECTOR_FOR_DELETE); } diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/SchemasTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/SchemasTests.java index b8139024e7b..8fa81c98b17 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/SchemasTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/SchemasTests.java @@ -3,8 +3,6 @@ import com.provectus.kafka.ui.api.model.CompatibilityLevel; import com.provectus.kafka.ui.base.BaseTest; import com.provectus.kafka.ui.models.Schema; -import com.provectus.kafka.ui.pages.MainPage; -import com.provectus.kafka.ui.pages.schema.SchemaDetails; import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.AutomationStatus; import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.Suite; import com.provectus.kafka.ui.utilities.qaseIoUtils.enums.Status; @@ -15,6 +13,8 @@ import java.util.ArrayList; import java.util.List; +import static com.provectus.kafka.ui.pages.NaviSideBar.SideMenuOption.SCHEMA_REGISTRY; +import static com.provectus.kafka.ui.settings.Source.CLUSTER_NAME; import static com.provectus.kafka.ui.utilities.FileUtils.fileToString; @TestInstance(TestInstance.Lifecycle.PER_CLASS) @@ -42,16 +42,22 @@ public void beforeAll() { @Order(1) void createSchemaAvro() { Schema schemaAvro = Schema.createSchemaAvro(); - mainPage.goTo() - .goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.SCHEMA_REGISTRY); - schemaRegistryList.clickCreateSchema() + naviSideBar + .openSideMenu(SCHEMA_REGISTRY); + schemaRegistryList + .waitUntilScreenReady() + .clickCreateSchema(); + schemaCreateForm .setSubjectName(schemaAvro.getName()) .setSchemaField(fileToString(schemaAvro.getValuePath())) .selectSchemaTypeFromDropdown(schemaAvro.getType()) - .clickSubmit() + .clickSubmit(); + schemaDetails + .waitUntilScreenReady(); + naviSideBar + .openSideMenu(SCHEMA_REGISTRY); + schemaRegistryList .waitUntilScreenReady(); - mainPage.goTo() - .goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.SCHEMA_REGISTRY); Assertions.assertTrue(schemaRegistryList.isSchemaVisible(schemaAvro.getName()),"isSchemaVisible()"); SCHEMA_LIST.add(schemaAvro); } @@ -64,17 +70,24 @@ void createSchemaAvro() { @Order(2) void updateSchemaAvro() { AVRO_API.setValuePath(System.getProperty("user.dir") + "/src/main/resources/testData/schema_avro_for_update.json"); - mainPage.goTo() - .goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.SCHEMA_REGISTRY); - schemaRegistryList.openSchema(AVRO_API.getName()) + naviSideBar + .openSideMenu(SCHEMA_REGISTRY); + schemaRegistryList + .waitUntilScreenReady() + .openSchema(AVRO_API.getName()); + schemaDetails .waitUntilScreenReady() .openEditSchema(); + schemaCreateForm + .waitUntilScreenReady(); Assertions.assertTrue(schemaCreateForm.isSchemaDropDownDisabled(),"isSchemaDropDownDisabled()"); - schemaCreateForm.selectCompatibilityLevelFromDropdown(CompatibilityLevel.CompatibilityEnum.NONE) + schemaCreateForm + .selectCompatibilityLevelFromDropdown(CompatibilityLevel.CompatibilityEnum.NONE) .setNewSchemaValue(fileToString(AVRO_API.getValuePath())) - .clickSubmit() + .clickSubmit(); + schemaDetails .waitUntilScreenReady(); - Assertions.assertEquals(CompatibilityLevel.CompatibilityEnum.NONE.toString(), new SchemaDetails().getCompatibility(), "getCompatibility()"); + Assertions.assertEquals(CompatibilityLevel.CompatibilityEnum.NONE.toString(), schemaDetails.getCompatibility(), "getCompatibility()"); } @DisplayName("should delete AVRO schema") @@ -84,11 +97,16 @@ void updateSchemaAvro() { @Test @Order(3) void deleteSchemaAvro() { - mainPage.goTo() - .goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.SCHEMA_REGISTRY); - schemaRegistryList.openSchema(AVRO_API.getName()) + naviSideBar + .openSideMenu(SCHEMA_REGISTRY); + schemaRegistryList + .waitUntilScreenReady() + .openSchema(AVRO_API.getName()); + schemaDetails .waitUntilScreenReady() .removeSchema(); + schemaRegistryList + .waitUntilScreenReady(); Assertions.assertFalse(schemaRegistryList.isSchemaVisible(AVRO_API.getName()),"isSchemaVisible()"); SCHEMA_LIST.remove(AVRO_API); } @@ -101,15 +119,22 @@ void deleteSchemaAvro() { @Order(4) void createSchemaJson() { Schema schemaJson = Schema.createSchemaJson(); - mainPage.goTo() - .goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.SCHEMA_REGISTRY); - schemaRegistryList.clickCreateSchema() + naviSideBar + .openSideMenu(SCHEMA_REGISTRY); + schemaRegistryList + .waitUntilScreenReady() + .clickCreateSchema(); + schemaCreateForm .setSubjectName(schemaJson.getName()) .setSchemaField(fileToString(schemaJson.getValuePath())) .selectSchemaTypeFromDropdown(schemaJson.getType()) - .clickSubmit() + .clickSubmit(); + schemaDetails + .waitUntilScreenReady(); + naviSideBar + .openSideMenu(SCHEMA_REGISTRY); + schemaRegistryList .waitUntilScreenReady(); - mainPage.goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.SCHEMA_REGISTRY); Assertions.assertTrue(schemaRegistryList.isSchemaVisible(schemaJson.getName()),"isSchemaVisible()"); SCHEMA_LIST.add(schemaJson); } @@ -121,11 +146,16 @@ void createSchemaJson() { @Test @Order(5) void deleteSchemaJson() { - mainPage.goTo() - .goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.SCHEMA_REGISTRY); - schemaRegistryList.openSchema(JSON_API.getName()) + naviSideBar + .openSideMenu(SCHEMA_REGISTRY); + schemaRegistryList + .waitUntilScreenReady() + .openSchema(JSON_API.getName()); + schemaDetails .waitUntilScreenReady() .removeSchema(); + schemaRegistryList + .waitUntilScreenReady(); Assertions.assertFalse(schemaRegistryList.isSchemaVisible(JSON_API.getName()),"isSchemaVisible()"); SCHEMA_LIST.remove(JSON_API); } @@ -138,15 +168,22 @@ void deleteSchemaJson() { @Order(6) void createSchemaProtobuf() { Schema schemaProtobuf = Schema.createSchemaProtobuf(); - mainPage.goTo() - .goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.SCHEMA_REGISTRY); - schemaRegistryList.clickCreateSchema() + naviSideBar + .openSideMenu(SCHEMA_REGISTRY); + schemaRegistryList + .waitUntilScreenReady() + .clickCreateSchema(); + schemaCreateForm .setSubjectName(schemaProtobuf.getName()) .setSchemaField(fileToString(schemaProtobuf.getValuePath())) .selectSchemaTypeFromDropdown(schemaProtobuf.getType()) - .clickSubmit() + .clickSubmit(); + schemaDetails + .waitUntilScreenReady(); + naviSideBar + .openSideMenu(SCHEMA_REGISTRY); + schemaRegistryList .waitUntilScreenReady(); - mainPage.goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.SCHEMA_REGISTRY); Assertions.assertTrue(schemaRegistryList.isSchemaVisible(schemaProtobuf.getName()),"isSchemaVisible()"); SCHEMA_LIST.add(schemaProtobuf); } @@ -158,11 +195,16 @@ void createSchemaProtobuf() { @Test @Order(7) void deleteSchemaProtobuf() { - mainPage.goTo() - .goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.SCHEMA_REGISTRY); - schemaRegistryList.openSchema(PROTOBUF_API.getName()) + naviSideBar + .openSideMenu(SCHEMA_REGISTRY); + schemaRegistryList + .waitUntilScreenReady() + .openSchema(PROTOBUF_API.getName()); + schemaDetails .waitUntilScreenReady() .removeSchema(); + schemaRegistryList + .waitUntilScreenReady(); Assertions.assertFalse(schemaRegistryList.isSchemaVisible(PROTOBUF_API.getName()),"isSchemaVisible()"); SCHEMA_LIST.remove(PROTOBUF_API); } diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java index 03628ecbfd3..378190c6302 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java @@ -2,7 +2,6 @@ import com.provectus.kafka.ui.base.BaseTest; import com.provectus.kafka.ui.models.Topic; -import com.provectus.kafka.ui.pages.MainPage; import com.provectus.kafka.ui.pages.topic.TopicDetails; import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.AutomationStatus; import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.Suite; @@ -14,6 +13,8 @@ import java.util.ArrayList; import java.util.List; +import static com.provectus.kafka.ui.pages.NaviSideBar.SideMenuOption.TOPICS; +import static com.provectus.kafka.ui.settings.Source.CLUSTER_NAME; import static com.provectus.kafka.ui.utilities.FileUtils.fileToString; @TestInstance(TestInstance.Lifecycle.PER_CLASS) @@ -44,15 +45,21 @@ public void beforeAll() { @Test public void createTopic() { Topic topicToCreate = new Topic().setName("new-topic"); - mainPage.goTo() - .goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.TOPICS); - topicsList.pressCreateNewTopic() + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .pressCreateNewTopic(); + topicCreateEditForm .setTopicName(topicToCreate.getName()) - .sendData() + .sendData(); + topicDetails + .waitUntilScreenReady(); + naviSideBar + .openSideMenu(TOPICS); + topicsList .waitUntilScreenReady(); - mainPage.goTo() - .goToSideMenu(CLUSTER_NAME, MainPage.SideMenuOptions.TOPICS); - Assertions.assertTrue(topicsList.isTopicVisible(topicToCreate.getName()),"isTopicVisible"); + Assertions.assertTrue(topicsList.isTopicVisible(topicToCreate.getName()), "isTopicVisible"); TOPIC_LIST.add(topicToCreate); } @@ -63,21 +70,29 @@ public void createTopic() { @CaseId(197) @Test public void updateTopic() { - topicsList.goTo(CLUSTER_NAME) - .waitUntilScreenReady(); - topicsList.openTopic(TOPIC_FOR_UPDATE.getName()) + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .openTopic(TOPIC_FOR_UPDATE.getName()); + topicDetails .waitUntilScreenReady() - .openEditSettings() + .openEditSettings(); + topicCreateEditForm .selectCleanupPolicy(TOPIC_FOR_UPDATE.getCompactPolicyValue()) .setMinInsyncReplicas(10) .setTimeToRetainDataInMs(TOPIC_FOR_UPDATE.getTimeToRetainData()) .setMaxSizeOnDiskInGB(TOPIC_FOR_UPDATE.getMaxSizeOnDisk()) .setMaxMessageBytes(TOPIC_FOR_UPDATE.getMaxMessageBytes()) - .sendData() - .waitUntilScreenReady(); - topicsList.goTo(CLUSTER_NAME) + .sendData(); + topicDetails .waitUntilScreenReady(); - topicsList.openTopic(TOPIC_FOR_UPDATE.getName()) + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .openTopic(TOPIC_FOR_UPDATE.getName()); + topicDetails .waitUntilScreenReady() .openEditSettings(); SoftAssertions softly = new SoftAssertions(); @@ -94,14 +109,19 @@ public void updateTopic() { @CaseId(207) @Test public void deleteTopic() { - topicsList.goTo(CLUSTER_NAME) + naviSideBar + .openSideMenu(TOPICS); + topicsList .waitUntilScreenReady() - .openTopic(TOPIC_FOR_DELETE.getName()) + .openTopic(TOPIC_FOR_DELETE.getName()); + topicDetails .waitUntilScreenReady() .deleteTopic(); - topicsList.goTo(CLUSTER_NAME) + naviSideBar + .openSideMenu(TOPICS); + topicsList .waitUntilScreenReady(); - Assertions.assertFalse(topicsList.isTopicVisible(TOPIC_FOR_DELETE.getName()),"isTopicVisible"); + Assertions.assertFalse(topicsList.isTopicVisible(TOPIC_FOR_DELETE.getName()), "isTopicVisible"); TOPIC_LIST.remove(TOPIC_FOR_DELETE); } @@ -111,15 +131,21 @@ public void deleteTopic() { @CaseId(222) @Test void produceMessage() { - topicsList.goTo(CLUSTER_NAME) + naviSideBar + .openSideMenu(TOPICS); + topicsList .waitUntilScreenReady() - .openTopic(TOPIC_FOR_UPDATE.getName()) + .openTopic(TOPIC_FOR_UPDATE.getName()); + topicDetails .waitUntilScreenReady() .openTopicMenu(TopicDetails.TopicMenu.MESSAGES) - .clickOnButton("Produce Message") + .clickOnButton("Produce Message"); + produceMessagePanel .setContentFiled(TOPIC_FOR_UPDATE.getMessageContent()) .setKeyField(TOPIC_FOR_UPDATE.getMessageKey()) .submitProduceMessage(); + topicDetails + .waitUntilScreenReady(); SoftAssertions softly = new SoftAssertions(); softly.assertThat(topicDetails.isKeyMessageVisible((TOPIC_FOR_UPDATE.getMessageKey()))).withFailMessage("isKeyMessageVisible()").isTrue(); softly.assertThat(topicDetails.isContentMessageVisible((TOPIC_FOR_UPDATE.getMessageContent()).trim())).withFailMessage("isContentMessageVisible()").isTrue();
test
val
2022-10-07T12:36:27
"2022-09-26T13:47:08Z"
VladSenyuta
train
provectus/kafka-ui/2684_2704
provectus/kafka-ui
provectus/kafka-ui/2684
provectus/kafka-ui/2704
[ "timestamp(timedelta=2.0, similarity=0.9443145359730044)", "connected" ]
0939d6140f5540bd421f19d61d862d363cff7456
6581ee605b7574ecdb64a69fc88fddb47dde53bd
[]
[ "Let's make dropdown item disabled instead of removing it from dom", "it is disabled for internal", "should be disabled", "```suggestion\r\n expect(screen.getByLabelText('Dropdown Toggle')).toBeDisabled();\r\n```", "```suggestion\r\n expect(screen.getByLabelText('Dropdown Toggle')).toBeDisabled();\r\n```", "remove this line to make experiment clear", "remove this line to make experiment clear", "we can completely remove this `it` case" ]
"2022-10-05T21:18:06Z"
[ "good first issue", "scope/frontend", "type/refactoring", "status/accepted" ]
Migrate Topic Overview to the new version of table
### Describe the solution you'd like Migrate `src/components/Topics/Topic/Overview/Overview.tsx` to new version of tables `src/components/common/NewTable/Table.tsx` ### Additional context Topics -> Topic
[ "kafka-ui-react-app/src/components/Topics/Topic/Overview/Overview.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Overview/__test__/Overview.spec.tsx" ]
[ "kafka-ui-react-app/src/components/Topics/Topic/Overview/ActionsCell.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Overview/Overview.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Overview/__test__/Overview.spec.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Overview/ActionsCell.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Overview/ActionsCell.tsx new file mode 100644 index 00000000000..a9f4eb15e8d --- /dev/null +++ b/kafka-ui-react-app/src/components/Topics/Topic/Overview/ActionsCell.tsx @@ -0,0 +1,35 @@ +import React from 'react'; +import { Partition } from 'generated-sources'; +import { CellContext } from '@tanstack/react-table'; +import { useAppDispatch } from 'lib/hooks/redux'; +import ClusterContext from 'components/contexts/ClusterContext'; +import { RouteParamsClusterTopic } from 'lib/paths'; +import useAppParams from 'lib/hooks/useAppParams'; +import { clearTopicMessages } from 'redux/reducers/topicMessages/topicMessagesSlice'; +import { Dropdown, DropdownItem } from 'components/common/Dropdown'; +import { useTopicDetails } from 'lib/hooks/api/topics'; + +const ActionsCell: React.FC<CellContext<Partition, unknown>> = ({ row }) => { + const { clusterName, topicName } = useAppParams<RouteParamsClusterTopic>(); + const { data } = useTopicDetails({ clusterName, topicName }); + const { isReadOnly } = React.useContext(ClusterContext); + const { partition } = row.original; + const dispatch = useAppDispatch(); + + const clearTopicMessagesHandler = async () => { + await dispatch( + clearTopicMessages({ clusterName, topicName, partitions: [partition] }) + ).unwrap(); + }; + const disabled = + data?.internal || isReadOnly || data?.cleanUpPolicy !== 'DELETE'; + return ( + <Dropdown disabled={disabled}> + <DropdownItem onClick={clearTopicMessagesHandler} danger> + Clear Messages + </DropdownItem> + </Dropdown> + ); +}; + +export default ActionsCell; diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Overview/Overview.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Overview/Overview.tsx index 2bd1c4a78a6..cada465548f 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Overview/Overview.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Overview/Overview.tsx @@ -1,25 +1,20 @@ import React from 'react'; import { Partition, Replica } from 'generated-sources'; -import ClusterContext from 'components/contexts/ClusterContext'; import BytesFormatted from 'components/common/BytesFormatted/BytesFormatted'; -import { Table } from 'components/common/table/Table/Table.styled'; -import TableHeaderCell from 'components/common/table/TableHeaderCell/TableHeaderCell'; +import Table from 'components/common/NewTable'; import * as Metrics from 'components/common/Metrics'; import { Tag } from 'components/common/Tag/Tag.styled'; -import { useAppDispatch } from 'lib/hooks/redux'; import { RouteParamsClusterTopic } from 'lib/paths'; import useAppParams from 'lib/hooks/useAppParams'; -import { Dropdown, DropdownItem } from 'components/common/Dropdown'; -import { clearTopicMessages } from 'redux/reducers/topicMessages/topicMessagesSlice'; import { useTopicDetails } from 'lib/hooks/api/topics'; +import { ColumnDef } from '@tanstack/react-table'; import * as S from './Overview.styled'; +import ActionsCell from './ActionsCell'; const Overview: React.FC = () => { const { clusterName, topicName } = useAppParams<RouteParamsClusterTopic>(); - const dispatch = useAppDispatch(); const { data } = useTopicDetails({ clusterName, topicName }); - const { isReadOnly } = React.useContext(ClusterContext); const messageCount = React.useMemo( () => @@ -28,7 +23,65 @@ const Overview: React.FC = () => { }, 0), [data] ); + const newData = React.useMemo(() => { + if (!data?.partitions) return []; + return data.partitions.map((items: Partition) => { + return { + ...items, + messageCount: items.offsetMax - items.offsetMin, + }; + }); + }, [data?.partitions]); + + const columns = React.useMemo<ColumnDef<Partition>[]>( + () => [ + { + header: 'Partition ID', + enableSorting: false, + accessorKey: 'partition', + }, + { + header: 'Replicas', + enableSorting: false, + + accessorKey: 'replicas', + cell: ({ getValue }) => { + const replicas = getValue<Partition['replicas']>(); + if (replicas === undefined || replicas.length === 0) { + return 0; + } + return replicas?.map(({ broker, leader }: Replica) => ( + <S.Replica + leader={leader} + key={broker} + title={leader ? 'Leader' : ''} + > + {broker} + </S.Replica> + )); + }, + }, + { + header: 'First Offset', + enableSorting: false, + accessorKey: 'offsetMin', + }, + { header: 'Next Offset', enableSorting: false, accessorKey: 'offsetMax' }, + { + header: 'Message Count', + enableSorting: false, + accessorKey: `messageCount`, + }, + { + header: '', + enableSorting: false, + accessorKey: 'actions', + cell: ActionsCell, + }, + ], + [] + ); return ( <> <Metrics.Wrapper> @@ -90,66 +143,12 @@ const Overview: React.FC = () => { </Metrics.Indicator> </Metrics.Section> </Metrics.Wrapper> - <Table isFullwidth> - <thead> - <tr> - <TableHeaderCell title="Partition ID" /> - <TableHeaderCell title="Replicas" /> - <TableHeaderCell title="First Offset" /> - <TableHeaderCell title="Next Offset" /> - <TableHeaderCell title="Message Count" /> - <TableHeaderCell title=" " /> - </tr> - </thead> - <tbody> - {data?.partitions?.map((partition: Partition) => ( - <tr key={`partition-list-item-key-${partition.partition}`}> - <td>{partition.partition}</td> - <td> - {partition.replicas?.map(({ broker, leader }: Replica) => ( - <S.Replica - leader={leader} - key={broker} - title={leader ? 'Leader' : ''} - > - {broker} - </S.Replica> - ))} - </td> - <td>{partition.offsetMin}</td> - <td>{partition.offsetMax}</td> - <td>{partition.offsetMax - partition.offsetMin}</td> - <td style={{ width: '5%' }}> - {!data?.internal && - !isReadOnly && - data?.cleanUpPolicy === 'DELETE' ? ( - <Dropdown> - <DropdownItem - onClick={() => - dispatch( - clearTopicMessages({ - clusterName, - topicName, - partitions: [partition.partition], - }) - ).unwrap() - } - danger - > - Clear Messages - </DropdownItem> - </Dropdown> - ) : null} - </td> - </tr> - ))} - {data?.partitions?.length === 0 && ( - <tr> - <td colSpan={10}>No Partitions found</td> - </tr> - )} - </tbody> - </Table> + <Table + columns={columns} + data={newData} + enableSorting + emptyMessage="No Partitions found " + /> </> ); }; diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Overview/__test__/Overview.spec.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Overview/__test__/Overview.spec.tsx index f49606747e4..e0efd0db9ec 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Overview/__test__/Overview.spec.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Overview/__test__/Overview.spec.tsx @@ -122,39 +122,29 @@ describe('Overview', () => { }); describe('when the table partition dropdown appearance', () => { - it('should check if the dropdown is not present when it is readOnly', () => { + it('should check if the dropdown is disabled when it is readOnly', () => { renderComponent( { - ...internalTopicPayload, - cleanUpPolicy: CleanUpPolicy.DELETE, + ...externalTopicPayload, }, { ...defaultContextValues, isReadOnly: true } ); - expect(screen.queryByText('Clear Messages')).not.toBeInTheDocument(); + expect(screen.getByLabelText('Dropdown Toggle')).toBeDisabled(); }); - it('should check if the dropdown is not present when it is internal', () => { + it('should check if the dropdown is disabled when it is internal', () => { renderComponent({ ...internalTopicPayload, - cleanUpPolicy: CleanUpPolicy.DELETE, }); - expect(screen.queryByText('Clear Messages')).not.toBeInTheDocument(); + expect(screen.getByLabelText('Dropdown Toggle')).toBeDisabled(); }); - it('should check if the dropdown is not present when cleanUpPolicy is not DELETE', () => { + it('should check if the dropdown is disabled when cleanUpPolicy is not DELETE', () => { renderComponent({ ...externalTopicPayload, cleanUpPolicy: CleanUpPolicy.COMPACT, }); - expect(screen.queryByText('Clear Messages')).not.toBeInTheDocument(); - }); - - it('should check if the dropdown action to be in visible', () => { - renderComponent({ - ...externalTopicPayload, - cleanUpPolicy: CleanUpPolicy.DELETE, - }); - expect(screen.getByText('Clear Messages')).toBeInTheDocument(); + expect(screen.getByLabelText('Dropdown Toggle')).toBeDisabled(); }); }); });
null
train
val
2022-10-10T17:36:53
"2022-09-30T10:55:07Z"
workshur
train
provectus/kafka-ui/2683_2708
provectus/kafka-ui
provectus/kafka-ui/2683
provectus/kafka-ui/2708
[ "keyword_pr_to_issue", "timestamp(timedelta=1.0, similarity=1.0)" ]
2b5dd270e22c6455c6a5999131dbf3ce3b770429
45aa4542ba6e502d8d627d444035df1634174a69
[]
[ "pls use fixtures" ]
"2022-10-06T13:09:20Z"
[ "scope/frontend", "type/refactoring", "status/accepted" ]
Migrate Schema details table to the new version of Table component
### Describe the solution you'd like Migrate `src/components/Schemas/Details/Details.tsx` to new version of tables `src/components/common/NewTable/Table.tsx` ### Additional context Schema Registry -> Schema
[ "kafka-ui-react-app/src/components/Schemas/Details/Details.tsx", "kafka-ui-react-app/src/components/Schemas/Details/SchemaVersion/SchemaVersion.styled.ts", "kafka-ui-react-app/src/components/Schemas/Details/SchemaVersion/SchemaVersion.tsx", "kafka-ui-react-app/src/components/Schemas/Details/__test__/Details.spec.tsx", "kafka-ui-react-app/src/components/Schemas/Details/__test__/SchemaVersion.spec.tsx" ]
[ "kafka-ui-react-app/src/components/Schemas/Details/Details.tsx", "kafka-ui-react-app/src/components/Schemas/Details/SchemaVersion/SchemaVersion.tsx", "kafka-ui-react-app/src/components/Schemas/Details/__test__/Details.spec.tsx", "kafka-ui-react-app/src/components/Schemas/Details/__test__/SchemaVersion.spec.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Schemas/Details/Details.tsx b/kafka-ui-react-app/src/components/Schemas/Details/Details.tsx index fb27336493a..8878728fa5d 100644 --- a/kafka-ui-react-app/src/components/Schemas/Details/Details.tsx +++ b/kafka-ui-react-app/src/components/Schemas/Details/Details.tsx @@ -10,8 +10,6 @@ import ClusterContext from 'components/contexts/ClusterContext'; import PageLoader from 'components/common/PageLoader/PageLoader'; import PageHeading from 'components/common/PageHeading/PageHeading'; import { Button } from 'components/common/Button/Button'; -import { Table } from 'components/common/table/Table/Table.styled'; -import TableHeaderCell from 'components/common/table/TableHeaderCell/TableHeaderCell'; import { useAppDispatch, useAppSelector } from 'lib/hooks/redux'; import { fetchLatestSchema, @@ -29,6 +27,7 @@ import { TableTitle } from 'components/common/table/TableTitle/TableTitle.styled import useAppParams from 'lib/hooks/useAppParams'; import { schemasApiClient } from 'lib/api'; import { Dropdown, DropdownItem } from 'components/common/Dropdown'; +import Table from 'components/common/NewTable'; import LatestVersionItem from './LatestVersion/LatestVersionItem'; import SchemaVersion from './SchemaVersion/SchemaVersion'; @@ -58,6 +57,15 @@ const Details: React.FC = () => { const isFetched = useAppSelector(getAreSchemaLatestFulfilled); const areVersionsFetched = useAppSelector(getAreSchemaVersionsFulfilled); + const columns = React.useMemo( + () => [ + { header: 'Version', accessorKey: 'version' }, + { header: 'ID', accessorKey: 'id' }, + { header: 'Type', accessorKey: 'schemaType' }, + ], + [] + ); + const deleteHandler = async () => { try { await schemasApiClient.deleteSchema({ @@ -118,26 +126,13 @@ const Details: React.FC = () => { <LatestVersionItem schema={schema} /> <TableTitle>Old versions</TableTitle> {areVersionsFetched ? ( - <Table isFullwidth> - <thead> - <tr> - <TableHeaderCell /> - <TableHeaderCell title="Version" /> - <TableHeaderCell title="ID" /> - <TableHeaderCell title="Type" /> - </tr> - </thead> - <tbody> - {versions.map((version) => ( - <SchemaVersion key={version.id} version={version} /> - ))} - {versions.length === 0 && ( - <tr> - <td colSpan={10}>No active Schema</td> - </tr> - )} - </tbody> - </Table> + <Table + columns={columns} + data={versions} + getRowCanExpand={() => true} + renderSubComponent={SchemaVersion} + enableSorting + /> ) : ( <PageLoader /> )} diff --git a/kafka-ui-react-app/src/components/Schemas/Details/SchemaVersion/SchemaVersion.styled.ts b/kafka-ui-react-app/src/components/Schemas/Details/SchemaVersion/SchemaVersion.styled.ts deleted file mode 100644 index fad9ee37849..00000000000 --- a/kafka-ui-react-app/src/components/Schemas/Details/SchemaVersion/SchemaVersion.styled.ts +++ /dev/null @@ -1,13 +0,0 @@ -import styled from 'styled-components'; - -export const Wrapper = styled.tr` - background-color: ${({ theme }) => theme.schema.backgroundColor.tr}; - & > td { - padding: 16px !important; - & > div { - background-color: ${({ theme }) => theme.schema.backgroundColor.div}; - border-radius: 8px; - padding: 24px; - } - } -`; diff --git a/kafka-ui-react-app/src/components/Schemas/Details/SchemaVersion/SchemaVersion.tsx b/kafka-ui-react-app/src/components/Schemas/Details/SchemaVersion/SchemaVersion.tsx index 472a6e544fd..fbba8c85010 100644 --- a/kafka-ui-react-app/src/components/Schemas/Details/SchemaVersion/SchemaVersion.tsx +++ b/kafka-ui-react-app/src/components/Schemas/Details/SchemaVersion/SchemaVersion.tsx @@ -1,41 +1,18 @@ import React from 'react'; -import { SchemaSubject } from 'generated-sources'; -import MessageToggleIcon from 'components/common/Icons/MessageToggleIcon'; -import IconButtonWrapper from 'components/common/Icons/IconButtonWrapper'; import EditorViewer from 'components/common/EditorViewer/EditorViewer'; +import { SchemaSubject } from 'generated-sources'; +import { Row } from '@tanstack/react-table'; -import * as S from './SchemaVersion.styled'; - -interface SchemaVersionProps { - version: SchemaSubject; +export interface Props { + row: Row<SchemaSubject>; } -const SchemaVersion: React.FC<SchemaVersionProps> = ({ - version: { version, id, schema, schemaType }, -}) => { - const [isOpen, setIsOpen] = React.useState(false); - const toggleIsOpen = () => setIsOpen(!isOpen); - +const SchemaVersion: React.FC<Props> = ({ row }) => { return ( - <> - <tr> - <td style={{ width: '3%' }}> - <IconButtonWrapper onClick={toggleIsOpen}> - <MessageToggleIcon isOpen={isOpen} /> - </IconButtonWrapper> - </td> - <td style={{ width: '6%' }}>{version}</td> - <td style={{ width: '6%' }}>{id}</td> - <td>{schemaType}</td> - </tr> - {isOpen && ( - <S.Wrapper> - <td colSpan={4}> - <EditorViewer data={schema} schemaType={schemaType} /> - </td> - </S.Wrapper> - )} - </> + <EditorViewer + data={row?.original?.schema} + schemaType={row?.original?.schemaType} + /> ); }; diff --git a/kafka-ui-react-app/src/components/Schemas/Details/__test__/Details.spec.tsx b/kafka-ui-react-app/src/components/Schemas/Details/__test__/Details.spec.tsx index a09788b162e..20f3870c7e5 100644 --- a/kafka-ui-react-app/src/components/Schemas/Details/__test__/Details.spec.tsx +++ b/kafka-ui-react-app/src/components/Schemas/Details/__test__/Details.spec.tsx @@ -22,6 +22,12 @@ const clusterName = 'testClusterName'; const schemasAPILatestUrl = `/api/clusters/${clusterName}/schemas/${schemaVersion.subject}/latest`; const schemasAPIVersionsUrl = `/api/clusters/${clusterName}/schemas/${schemaVersion.subject}/versions`; +const mockHistoryPush = jest.fn(); +jest.mock('react-router-dom', () => ({ + ...jest.requireActual('react-router-dom'), + useNavigate: () => mockHistoryPush, +})); + const renderComponent = ( initialState: RootState['schemas'] = schemasInitialState, context: ContextProps = contextInitialValue @@ -62,6 +68,27 @@ describe('Details', () => { }); }); + it('handles [Delete schema] click', async () => { + const deleteSchemaMock = fetchMock.deleteOnce( + `/api/clusters/${clusterName}/schemas/${schemaVersion.subject}`, + 200 + ); + + await act(() => { + renderComponent(); + }); + + try { + expect(deleteSchemaMock.called()).toBeTruthy(); + expect(mockHistoryPush).toHaveBeenCalledTimes(1); + expect(mockHistoryPush).toHaveBeenCalledWith( + clusterSchemaPath(clusterName) + ); + } catch (e) { + expect(deleteSchemaMock.called()).toBeTruthy(); + } + }); + it('renders pageloader', () => { expect(screen.getByRole('progressbar')).toBeInTheDocument(); expect(screen.queryByText(schemaVersion.subject)).not.toBeInTheDocument(); @@ -153,7 +180,6 @@ describe('Details', () => { // seems like incorrect behaviour it('renders versions table with 0 items', () => { expect(screen.getByRole('table')).toBeInTheDocument(); - expect(screen.getByText('No active Schema')).toBeInTheDocument(); }); }); }); diff --git a/kafka-ui-react-app/src/components/Schemas/Details/__test__/SchemaVersion.spec.tsx b/kafka-ui-react-app/src/components/Schemas/Details/__test__/SchemaVersion.spec.tsx index ccd50cd957f..2ec3b001a94 100644 --- a/kafka-ui-react-app/src/components/Schemas/Details/__test__/SchemaVersion.spec.tsx +++ b/kafka-ui-react-app/src/components/Schemas/Details/__test__/SchemaVersion.spec.tsx @@ -1,24 +1,25 @@ import React from 'react'; import SchemaVersion from 'components/Schemas/Details/SchemaVersion/SchemaVersion'; import { render } from 'lib/testHelpers'; -import { screen } from '@testing-library/react'; -import userEvent from '@testing-library/user-event'; +import { SchemaSubject } from 'generated-sources'; +import { Row } from '@tanstack/react-table'; -import { versions } from './fixtures'; +import { jsonSchema } from './fixtures'; -const component = ( - <table> - <tbody> - <SchemaVersion version={versions[0]} /> - </tbody> - </table> -); +export interface Props { + row: Row<SchemaSubject>; +} + +const renderComponent = () => { + const row = { + original: jsonSchema, + }; + + return render(<SchemaVersion row={row as Row<SchemaSubject>} />); +}; describe('SchemaVersion', () => { it('renders versions', () => { - render(component); - expect(screen.getAllByRole('cell')).toHaveLength(4); - expect(screen.queryByTestId('json-viewer')).not.toBeInTheDocument(); - userEvent.click(screen.getByRole('button')); + renderComponent(); }); });
null
val
val
2022-10-14T11:16:29
"2022-09-30T10:51:02Z"
workshur
train
provectus/kafka-ui/2681_2717
provectus/kafka-ui
provectus/kafka-ui/2681
provectus/kafka-ui/2717
[ "timestamp(timedelta=1.0, similarity=0.9453779959826046)", "connected" ]
0939d6140f5540bd421f19d61d862d363cff7456
15e7fc42d209c8f5334721798a92da0fc7340965
[]
[]
"2022-10-10T10:33:56Z"
[ "scope/frontend", "type/refactoring", "status/accepted" ]
Migrate Topic Consumers page to new version of table component
### Describe the solution you'd like Migrate `src/components/Topics/Topic/ConsumerGroups/TopicConsumerGroups.tsx` to new version of tables `src/components/common/NewTable/Table.tsx` ### Additional context Topics -> Topic -> Consumers
[ "kafka-ui-react-app/src/components/Topics/Topic/ConsumerGroups/TopicConsumerGroups.tsx" ]
[ "kafka-ui-react-app/src/components/Topics/Topic/ConsumerGroups/TopicConsumerGroups.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/Topic/ConsumerGroups/TopicConsumerGroups.tsx b/kafka-ui-react-app/src/components/Topics/Topic/ConsumerGroups/TopicConsumerGroups.tsx index 0fc72fd82e8..69948fd66d4 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/ConsumerGroups/TopicConsumerGroups.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/ConsumerGroups/TopicConsumerGroups.tsx @@ -1,65 +1,72 @@ import React from 'react'; -import { Link } from 'react-router-dom'; import { clusterConsumerGroupsPath, RouteParamsClusterTopic } from 'lib/paths'; -import { Table } from 'components/common/table/Table/Table.styled'; -import TableHeaderCell from 'components/common/table/TableHeaderCell/TableHeaderCell'; -import { Tag } from 'components/common/Tag/Tag.styled'; -import { TableKeyLink } from 'components/common/table/Table/TableKeyLink.styled'; -import getTagColor from 'components/common/Tag/getTagColor'; +import { ConsumerGroup } from 'generated-sources'; import useAppParams from 'lib/hooks/useAppParams'; import { useTopicConsumerGroups } from 'lib/hooks/api/topics'; +import { ColumnDef } from '@tanstack/react-table'; +import Table, { LinkCell, TagCell } from 'components/common/NewTable'; const TopicConsumerGroups: React.FC = () => { const { clusterName, topicName } = useAppParams<RouteParamsClusterTopic>(); - const { data: consumerGroups } = useTopicConsumerGroups({ + const { data: consumerGroups = [] } = useTopicConsumerGroups({ clusterName, topicName, }); + const columns = React.useMemo<ColumnDef<ConsumerGroup>[]>( + () => [ + { + header: 'Consumer Group ID', + accessorKey: 'groupId', + enableSorting: false, + // eslint-disable-next-line react/no-unstable-nested-components + cell: ({ row }) => ( + <LinkCell + value={row.original.groupId} + to={`${clusterConsumerGroupsPath(clusterName)}/${ + row.original.groupId + }`} + /> + ), + }, + { + header: 'Active Consumers', + accessorKey: 'members', + enableSorting: false, + }, + { + header: 'Messages Behind', + accessorKey: 'messagesBehind', + enableSorting: false, + }, + { + header: 'Coordinator', + accessorKey: 'coordinator', + enableSorting: false, + cell: ({ getValue }) => { + const coordinator = getValue<ConsumerGroup['coordinator']>(); + if (coordinator === undefined) { + return 0; + } + return coordinator.id; + }, + }, + { + header: 'State', + accessorKey: 'state', + enableSorting: false, + cell: TagCell, + }, + ], + [] + ); return ( - <Table isFullwidth> - <thead> - <tr> - <TableHeaderCell title="Consumer Group ID" /> - <TableHeaderCell title="Active Consumers" /> - <TableHeaderCell title="Messages Behind" /> - <TableHeaderCell title="Coordinator" /> - <TableHeaderCell title="State" /> - </tr> - </thead> - <tbody> - {consumerGroups?.map((consumer) => ( - <tr key={consumer.groupId}> - <TableKeyLink> - <Link - to={`${clusterConsumerGroupsPath(clusterName)}/${ - consumer.groupId - }`} - > - {consumer.groupId} - </Link> - </TableKeyLink> - <td>{consumer.members}</td> - <td>{consumer.messagesBehind}</td> - <td>{consumer.coordinator?.id}</td> - <td> - {consumer.state && ( - <Tag color={getTagColor(consumer.state)}>{`${consumer.state - .charAt(0) - .toUpperCase()}${consumer.state - .slice(1) - .toLowerCase()}`}</Tag> - )} - </td> - </tr> - ))} - {(!consumerGroups || consumerGroups.length === 0) && ( - <tr> - <td colSpan={10}>No active consumer groups</td> - </tr> - )} - </tbody> - </Table> + <Table + columns={columns} + data={consumerGroups} + enableSorting + emptyMessage="No active consumer groups" + /> ); };
null
val
val
2022-10-10T17:36:53
"2022-09-30T10:42:01Z"
workshur
train
provectus/kafka-ui/2185_2737
provectus/kafka-ui
provectus/kafka-ui/2185
provectus/kafka-ui/2737
[ "connected", "timestamp(timedelta=1.0, similarity=0.8866299837009638)" ]
0939d6140f5540bd421f19d61d862d363cff7456
19e38fb1bfe31d369bb9cd9b9eed5ddcbcef8b92
[ "Hello there SergKam! πŸ‘‹\n\nThank you and congratulations πŸŽ‰ for opening your very first issue in this project! πŸ’–\n\nIn case you want to claim this issue, please comment down below! We will try to get back to you as soon as we can. πŸ‘€", "Hey, thanks for reaching out.\r\n\r\nThere's your problem in your logs:\r\n`Caused by: java.lang.IllegalArgumentException: JAAS config entry not terminated by semi-colon`\r\nHere's an [example](https://github.com/provectus/kafka-ui/blob/master/documentation/compose/kafka-ui-sasl.yaml#L20)", "Unfortunately still not working. now the error\r\n```\r\nCaused by: org.apache.kafka.common.errors.InvalidRequestException: Amazon MSK Serverless doesn't support fetching broker or broker-logger configurations.\r\n```\r\n", "Okay, thanks. That's a totally different issue, it seems like we have to adjust some code to support _serverless_ MSK.", "is it supported now?", "> is it supported now?\r\n\r\nnot yet :(", "When is the plan to release version 0.5 @Haarolean ?\r\n", "> When is the plan to release version 0.5 @Haarolean ?\r\n\r\n@advissor as soon as roadmap features are done. I'm not sure this will be a part of 0.5, but it has a high priority on our list as well.", "Hello @SergKam , can you pls attach full log for latest exception?", "> Hello @SergKam , can you pls attach full log for latest exception?\r\n\r\n@iliax No, sorry. It was just an experiment, we decided not to use AWS MSK and switch to Confluent instead, so I deleted everything.\r\nBut you can try to set it up yourself using AWS MSK guide https://docs.aws.amazon.com/msk/latest/developerguide/serverless-getting-started.html", "> > Hello @SergKam , can you pls attach full log for latest exception?\r\n> \r\n> @iliax No, sorry. It was just an experiment, we decided not to use AWS MSK and switch to Confluent instead, so I deleted everything. But you can try to set it up yourself using AWS MSK guide https://docs.aws.amazon.com/msk/latest/developerguide/serverless-getting-started.html\r\n\r\nok, we will create cluster and check how it goes, thank you", "Please note that we need some frontend changes as well (#2739), hopefully will get them live soon as well.", "Hai team added msk cluster not getting data,can you please help here what i did mistake , geeting same error as above " ]
[]
"2022-10-12T19:49:25Z"
[ "scope/backend", "status/accepted", "type/feature" ]
Support AWS MSK Serverless
**Describe the bug** The kafka-ui is working but the cluster is shown as "offline" and without topics. **Set up** I'm trying to use kafka-ui vv0.4.0(521ba0c) as a frontend for AWS MSK Serverless cluster mostly to create and manage topics. I started a EC2 instance that uses IAM role with full access to the MSK cluster. Basically,I just follow the tutorial https://docs.aws.amazon.com/msk/latest/developerguide/serverless-getting-started.html I tested the setup with a standard kafka-topic.sh tools to make sure the cluster is accessible from the instance and IAM role allows all operations with the cluster. then I downloaded latest kafka-ui jar from github, and created application.yaml ``` logging.level.root: DEBUG kafka: clusters: - name: local bootstrapServers: boot-ivd3n3is.c1.kafka-serverless.eu-central-1.amazonaws.com:9098 properties: security.protocol: SASL_SSL sasl.mechanism: AWS_MSK_IAM sasl.client.callback.handler.class: software.amazon.msk.auth.iam.IAMClientCallbackHandler sasl.jaas.config: software.amazon.msk.auth.iam.IAMLoginModule required awsDebugCreds=true ``` and started it `java -jar kafka-ui-api-v0.4.0.jar` The web server is starting and working, but with no connection to Kafka. in console, I see the error ``` 2022-06-18 01:47:16,045 DEBUG [parallel-1] o.a.k.c.a.i.AdminMetadataManager: [AdminClient clientId=adminclient-84] Setting bootstrap cluster metadata Cluster(id = null, nodes = [boot-ivd3n3is.c1.kafka-serverless.eu-central-1.amazonaws.com:9098 (id: -1 rack: null)], partitions = [], controller = null). 2022-06-18 01:47:16,045 INFO [parallel-1] o.a.k.c.m.Metrics: Metrics scheduler closed 2022-06-18 01:47:16,045 INFO [parallel-1] o.a.k.c.m.Metrics: Closing reporter org.apache.kafka.common.metrics.JmxReporter 2022-06-18 01:47:16,045 INFO [parallel-1] o.a.k.c.m.Metrics: Metrics reporters closed 2022-06-18 01:47:16,045 ERROR [parallel-1] c.p.k.u.s.MetricsService: Failed to collect cluster local info java.lang.IllegalStateException: Error while creating AdminClient for Cluster local ``` **Expected behavior** **Screenshots** ![image](https://user-images.githubusercontent.com/3240529/174418764-33dbf007-ca53-4bfd-97b9-fa3243e13494.png) **Additional context** java 18.0.1.1 2022-04-22 Amazon linux 5.10 118-111.515.amzn2.x86_64 [log.txt.gz](https://github.com/provectus/kafka-ui/files/8932807/log.txt.gz)
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/exception/TopicMetadataException.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/BrokerService.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/FeatureService.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/TopicsService.java", "kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml" ]
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/exception/TopicMetadataException.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/BrokerService.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/FeatureService.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/TopicsService.java", "kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml" ]
[]
diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/exception/TopicMetadataException.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/exception/TopicMetadataException.java index 7ccceefe614..a659f94f971 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/exception/TopicMetadataException.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/exception/TopicMetadataException.java @@ -6,6 +6,10 @@ public TopicMetadataException(String message) { super(message); } + public TopicMetadataException(String message, Throwable cause) { + super(message, cause); + } + @Override public ErrorCode getErrorCode() { return ErrorCode.INVALID_ENTITY_STATE; diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/BrokerService.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/BrokerService.java index cd6ad93ba07..9c258fa8de8 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/BrokerService.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/BrokerService.java @@ -47,10 +47,7 @@ private Mono<Map<Integer, List<ConfigEntry>>> loadBrokersConfig( private Mono<List<ConfigEntry>> loadBrokersConfig( KafkaCluster cluster, Integer brokerId) { return loadBrokersConfig(cluster, Collections.singletonList(brokerId)) - .map(map -> map.values().stream() - .findFirst() - .orElseThrow(() -> new NotFoundException( - String.format("Config for broker %s not found", brokerId)))); + .map(map -> map.values().stream().findFirst().orElse(List.of())); } private Flux<InternalBrokerConfig> getBrokersConfig(KafkaCluster cluster, Integer brokerId) { @@ -81,13 +78,6 @@ public Flux<BrokerDTO> getBrokers(KafkaCluster cluster) { .flatMapMany(Flux::fromIterable); } - public Mono<Node> getController(KafkaCluster cluster) { - return adminClientService - .get(cluster) - .flatMap(ReactiveAdminClient::describeCluster) - .map(ReactiveAdminClient.ClusterDescription::getController); - } - public Mono<Void> updateBrokerLogDir(KafkaCluster cluster, Integer broker, BrokerLogdirUpdateDTO brokerLogDir) { diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/FeatureService.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/FeatureService.java index 24a824fb168..9097f2b25e7 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/FeatureService.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/FeatureService.java @@ -60,6 +60,6 @@ private Mono<Boolean> isTopicDeletionEnabled(KafkaCluster cluster, Node controll .filter(e -> e.name().equals(DELETE_TOPIC_ENABLED_SERVER_PROPERTY)) .map(e -> Boolean.parseBoolean(e.value())) .findFirst() - .orElse(false)); + .orElse(true)); } } diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java index 0556b3a0461..c08b839410e 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java @@ -18,6 +18,7 @@ import java.util.Map; import java.util.Optional; import java.util.Set; +import java.util.concurrent.CompletionException; import java.util.concurrent.ConcurrentHashMap; import java.util.concurrent.ExecutionException; import java.util.concurrent.atomic.AtomicInteger; @@ -54,6 +55,7 @@ import org.apache.kafka.common.config.ConfigResource; import org.apache.kafka.common.errors.GroupIdNotFoundException; import org.apache.kafka.common.errors.GroupNotEmptyException; +import org.apache.kafka.common.errors.InvalidRequestException; import org.apache.kafka.common.errors.UnknownTopicOrPartitionException; import org.apache.kafka.common.requests.DescribeLogDirsResponse; import reactor.core.publisher.Mono; @@ -81,7 +83,7 @@ public static class ClusterDescription { } public static Mono<ReactiveAdminClient> create(AdminClient adminClient) { - return getClusterVersionImpl(adminClient) + return getClusterVersion(adminClient) .map(ver -> new ReactiveAdminClient( adminClient, @@ -105,7 +107,13 @@ private static SupportedFeature getSupportedUpdateFeatureForVersion(String versi private static <T> Mono<T> toMono(KafkaFuture<T> future) { return Mono.<T>create(sink -> future.whenComplete((res, ex) -> { if (ex != null) { - sink.error(ex); + // KafkaFuture doc is unclear about what exception wrapper will be used + // (from docs it should be ExecutionException, be we actually see CompletionException, so checking both + if (ex instanceof CompletionException || ex instanceof ExecutionException) { + sink.error(ex.getCause()); //unwrapping exception + } else { + sink.error(ex); + } } else { sink.success(res); } @@ -166,17 +174,29 @@ private Mono<Map<String, List<ConfigEntry>>> getTopicsConfigImpl(Collection<Stri c -> List.copyOf(c.getValue().entries())))); } - public Mono<Map<Integer, List<ConfigEntry>>> loadBrokersConfig(List<Integer> brokerIds) { + private static Mono<Map<Integer, List<ConfigEntry>>> loadBrokersConfig(AdminClient client, List<Integer> brokerIds) { List<ConfigResource> resources = brokerIds.stream() .map(brokerId -> new ConfigResource(ConfigResource.Type.BROKER, Integer.toString(brokerId))) .collect(toList()); return toMono(client.describeConfigs(resources).all()) + .doOnError(InvalidRequestException.class, + th -> log.trace("Error while getting broker {} configs", brokerIds, th)) + // some kafka backends (like MSK serverless) do not support broker's configs retrieval, + // in that case InvalidRequestException will be thrown + .onErrorResume(InvalidRequestException.class, th -> Mono.just(Map.of())) .map(config -> config.entrySet().stream() .collect(toMap( c -> Integer.valueOf(c.getKey().name()), c -> new ArrayList<>(c.getValue().entries())))); } + /** + * Return per-broker configs or empty map if broker's configs retrieval not supported. + */ + public Mono<Map<Integer, List<ConfigEntry>>> loadBrokersConfig(List<Integer> brokerIds) { + return loadBrokersConfig(client, brokerIds); + } + public Mono<Map<String, TopicDescription>> describeTopics() { return listTopics(true).flatMap(this::describeTopics); } @@ -280,20 +300,16 @@ public Mono<ClusterDescription> describeCluster() { })); } - private static Mono<String> getClusterVersionImpl(AdminClient client) { - return toMono(client.describeCluster().controller()).flatMap(controller -> - toMono(client.describeConfigs( - List.of(new ConfigResource( - ConfigResource.Type.BROKER, String.valueOf(controller.id())))) - .all() - .thenApply(configs -> - configs.values().stream() - .map(Config::entries) - .flatMap(Collection::stream) - .filter(entry -> entry.name().contains("inter.broker.protocol.version")) - .findFirst().map(ConfigEntry::value) - .orElse("1.0-UNKNOWN") - ))); + private static Mono<String> getClusterVersion(AdminClient client) { + return toMono(client.describeCluster().controller()) + .flatMap(controller -> loadBrokersConfig(client, List.of(controller.id()))) + .map(configs -> configs.values().stream() + .flatMap(Collection::stream) + .filter(entry -> entry.name().contains("inter.broker.protocol.version")) + .findFirst() + .map(ConfigEntry::value) + .orElse("1.0-UNKNOWN") + ); } public Mono<Void> deleteConsumerGroups(Collection<String> groupIds) { @@ -306,10 +322,14 @@ public Mono<Void> deleteConsumerGroups(Collection<String> groupIds) { public Mono<Void> createTopic(String name, int numPartitions, - short replicationFactor, + @Nullable Integer replicationFactor, Map<String, String> configs) { - return toMono(client.createTopics( - List.of(new NewTopic(name, numPartitions, replicationFactor).configs(configs))).all()); + var newTopic = new NewTopic( + name, + Optional.of(numPartitions), + Optional.ofNullable(replicationFactor).map(Integer::shortValue) + ).configs(configs); + return toMono(client.createTopics(List.of(newTopic)).all()); } public Mono<Void> alterPartitionReassignments( diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/TopicsService.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/TopicsService.java index 1790a3005c2..2dffe0de75e 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/TopicsService.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/TopicsService.java @@ -170,11 +170,11 @@ private Mono<InternalTopic> createTopic(KafkaCluster c, ReactiveAdminClient admi adminClient.createTopic( topicData.getName(), topicData.getPartitions(), - topicData.getReplicationFactor().shortValue(), + topicData.getReplicationFactor(), topicData.getConfigs() ).thenReturn(topicData) ) - .onErrorResume(t -> Mono.error(new TopicMetadataException(t.getMessage()))) + .onErrorMap(t -> new TopicMetadataException(t.getMessage(), t)) .flatMap(topicData -> loadTopicAfterCreation(c, topicData.getName())); } @@ -194,7 +194,7 @@ public Mono<InternalTopic> recreateTopic(KafkaCluster cluster, String topicName) ac.createTopic( topic.getName(), topic.getPartitionCount(), - (short) topic.getReplicationFactor(), + topic.getReplicationFactor(), topic.getTopicConfigs() .stream() .collect(Collectors.toMap(InternalTopicConfig::getName, @@ -430,7 +430,7 @@ public Mono<InternalTopic> cloneTopic( ac.createTopic( newTopicName, topic.getPartitionCount(), - (short) topic.getReplicationFactor(), + topic.getReplicationFactor(), topic.getTopicConfigs() .stream() .collect(Collectors diff --git a/kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml b/kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml index 68b8afa65a3..af6c6c3e6ae 100644 --- a/kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml +++ b/kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml @@ -2261,7 +2261,6 @@ components: required: - name - partitions - - replicationFactor TopicUpdate: type: object
null
train
val
2022-10-10T17:36:53
"2022-06-18T02:23:43Z"
SergKam
train
provectus/kafka-ui/2621_2744
provectus/kafka-ui
provectus/kafka-ui/2621
provectus/kafka-ui/2744
[ "timestamp(timedelta=1.0, similarity=0.9043015474243244)", "connected" ]
63b6bb3fa10dd33b069c15151bf706a313970550
dc1785e3380a8db62b775c6cb88653babed46b58
[]
[ "there is no need in useCallback" ]
"2022-10-13T20:51:45Z"
[ "type/bug", "good first issue", "scope/frontend", "status/accepted", "status/confirmed", "hacktoberfest" ]
[UI] KSQL: Empty field in Stream properties causes errors
<!-- Don't forget to check for existing issues/discussions regarding your proposal. We might already have it. https://github.com/provectus/kafka-ui/issues https://github.com/provectus/kafka-ui/discussions --> **Describe the bug** When having empty field in stream properties "underfined" property is sent to backend. **Set up** kafka-ui master **Steps to Reproduce** <!-- We'd like you to provide an example setup (via docker-compose, helm, etc.) to reproduce the problem, especially with a complex setups. --> Steps to reproduce the behavior: 1. go to KSQL -> Execute Sql query 2. enter some query like `select * from GH_PRS_COUNTS_KSQL_TABLE_W;` and set stream property like `ksql.query.pull.table.scan.enabled = true`. 3. Press "Add Stream Property" button to create new empty field 4. Press Execute => ```{"ksql":"select * from GH_PRS_COUNTS_KSQL_TABLE_W;","streamsProperties":{"ksql.query.pull.table.scan.enabled":"true","undefined":""}}``` is sent to backend, which is not right **Expected behavior** We only sending properties that were directly set. ### NOTE: Within this ticket lets also prohibit adding more that one empty Stream Property input field - this looks very strange in UI when you press to "Add Stream Property" multiple times... **Screenshots** <!-- (If applicable, add screenshots to help explain your problem) --> <img width="1693" alt="Screenshot 2022-09-21 at 13 29 25" src="https://user-images.githubusercontent.com/702205/191471724-a5909ae1-380e-463b-9792-e5a95a9d3ada.png"> **Additional context** <!-- (Add any other context about the problem here) -->
[ "kafka-ui-react-app/src/components/KsqlDb/Query/Query.tsx", "kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/QueryForm.tsx", "kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/__test__/QueryForm.spec.tsx" ]
[ "kafka-ui-react-app/src/components/KsqlDb/Query/Query.tsx", "kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/QueryForm.tsx", "kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/__test__/QueryForm.spec.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/KsqlDb/Query/Query.tsx b/kafka-ui-react-app/src/components/KsqlDb/Query/Query.tsx index efda3932175..7e875748636 100644 --- a/kafka-ui-react-app/src/components/KsqlDb/Query/Query.tsx +++ b/kafka-ui-react-app/src/components/KsqlDb/Query/Query.tsx @@ -167,7 +167,10 @@ const Query: FC = () => { const submitHandler = useCallback( (values: FormValues) => { - const streamsProperties = values.streamsProperties.reduce( + const filteredProperties = values.streamsProperties.filter( + (property) => property.key != null + ); + const streamsProperties = filteredProperties.reduce( (acc, current) => ({ ...acc, [current.key as keyof string]: current.value, diff --git a/kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/QueryForm.tsx b/kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/QueryForm.tsx index ee78877d225..9965ea5189a 100644 --- a/kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/QueryForm.tsx +++ b/kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/QueryForm.tsx @@ -1,4 +1,4 @@ -import React from 'react'; +import React, { useCallback } from 'react'; import { FormError } from 'components/common/Input/Input.styled'; import { ErrorMessage } from '@hookform/error-message'; import { useForm, Controller, useFieldArray } from 'react-hook-form'; @@ -46,6 +46,7 @@ const QueryForm: React.FC<Props> = ({ const { handleSubmit, setValue, + getValues, control, formState: { errors }, } = useForm<FormValues>({ @@ -64,6 +65,16 @@ const QueryForm: React.FC<Props> = ({ name: 'streamsProperties', }); + const handleAddNewProperty = useCallback(() => { + if ( + getValues().streamsProperties.every((prop) => { + return prop.key; + }) + ) { + append({ key: '', value: '' }); + } + }, []); + return ( <S.QueryWrapper> <form onSubmit={handleSubmit(submitHandler)}> @@ -164,7 +175,7 @@ const QueryForm: React.FC<Props> = ({ type="button" buttonSize="M" buttonType="secondary" - onClick={() => append({ key: '', value: '' })} + onClick={handleAddNewProperty} > <PlusIcon /> Add Stream Property diff --git a/kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/__test__/QueryForm.spec.tsx b/kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/__test__/QueryForm.spec.tsx index 12a61a5fa5f..0b8aa60b78a 100644 --- a/kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/__test__/QueryForm.spec.tsx +++ b/kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/__test__/QueryForm.spec.tsx @@ -157,7 +157,7 @@ describe('QueryForm', () => { expect(submitFn.mock.calls.length).toBe(1); }); - it('add new property', async () => { + it('adds new property', async () => { renderComponent({ fetching: false, hasResults: false, @@ -166,6 +166,8 @@ describe('QueryForm', () => { submitHandler: jest.fn(), }); + const textbox = screen.getByLabelText('key'); + await userEvent.type(textbox, 'prop_name'); await act(() => { userEvent.click( screen.getByRole('button', { name: 'Add Stream Property' }) @@ -174,6 +176,23 @@ describe('QueryForm', () => { expect(screen.getAllByRole('textbox', { name: 'key' }).length).toEqual(2); }); + it("doesn't add new property", async () => { + renderComponent({ + fetching: false, + hasResults: false, + handleClearResults: jest.fn(), + handleSSECancel: jest.fn(), + submitHandler: jest.fn(), + }); + + await act(() => { + userEvent.click( + screen.getByRole('button', { name: 'Add Stream Property' }) + ); + }); + expect(screen.getAllByRole('textbox', { name: 'key' }).length).toEqual(1); + }); + it('delete stream property', async () => { await renderComponent({ fetching: false, @@ -184,6 +203,7 @@ describe('QueryForm', () => { }); await act(() => { + userEvent.paste(screen.getByRole('textbox', { name: 'key' }), 'test'); userEvent.click( screen.getByRole('button', { name: 'Add Stream Property' }) );
null
train
val
2022-10-17T14:24:14
"2022-09-21T09:41:53Z"
iliax
train
provectus/kafka-ui/2748_2749
provectus/kafka-ui
provectus/kafka-ui/2748
provectus/kafka-ui/2749
[ "keyword_pr_to_issue" ]
5d80b3837137d4c3c4ee249f2038435a68c4191c
c4bbbbe3da06779907a57b05354bcfcb9919ce81
[ "Hey, thanks for reaching out.\r\n\r\nWe'll be glad to see a PR addressing this issue.", "Hello there sherifkayad! πŸ‘‹\n\nThank you and congratulations πŸŽ‰ for opening your very first issue in this project! πŸ’–\n\nIn case you want to claim this issue, please comment down below! We will try to get back to you as soon as we can. πŸ‘€", "@Haarolean PR added to this issue ;) Looking forward." ]
[]
"2022-10-14T09:03:41Z"
[ "type/enhancement", "good first issue", "status/accepted", "scope/k8s" ]
No custom `annotations` on the level of the `Deployment` metadata
<!-- Don't forget to check for existing issues/discussions regarding your proposal. We might already have it. https://github.com/provectus/kafka-ui/issues https://github.com/provectus/kafka-ui/discussions --> **Describe the bug** <!--(A clear and concise description of what the bug is.)--> Not really a bug, however, a missing small feature. Deployment object `annotations` should be given as a configuration parameter to the users of the Helm Charts. A PR can follow to work on that. **Set up** <!-- How do you run the app? Please provide as much info as possible: 1. App version (docker image version or check commit hash in the top left corner in UI) 2. Helm chart version, if you use one 3. Any IAAC configs We might close the issue without further explanation if you don't provide such information. --> Using Helm **Steps to Reproduce** Steps to reproduce the behavior: **Expected behavior** <!-- (A clear and concise description of what you expected to happen) --> **Screenshots** <!-- (If applicable, add screenshots to help explain your problem) --> **Additional context** <!-- (Add any other context about the problem here) -->
[ "charts/kafka-ui/Chart.yaml", "charts/kafka-ui/templates/deployment.yaml", "charts/kafka-ui/values.yaml" ]
[ "charts/kafka-ui/Chart.yaml", "charts/kafka-ui/templates/deployment.yaml", "charts/kafka-ui/values.yaml" ]
[]
diff --git a/charts/kafka-ui/Chart.yaml b/charts/kafka-ui/Chart.yaml index 1ecf0eeecea..2ae296b4ee6 100644 --- a/charts/kafka-ui/Chart.yaml +++ b/charts/kafka-ui/Chart.yaml @@ -2,6 +2,6 @@ apiVersion: v2 name: kafka-ui description: A Helm chart for kafka-UI type: application -version: 0.4.5 +version: 0.4.6 appVersion: v0.4.0 icon: https://github.com/provectus/kafka-ui/raw/master/documentation/images/kafka-ui-logo.png diff --git a/charts/kafka-ui/templates/deployment.yaml b/charts/kafka-ui/templates/deployment.yaml index 51703c7ed30..a253fc6df51 100644 --- a/charts/kafka-ui/templates/deployment.yaml +++ b/charts/kafka-ui/templates/deployment.yaml @@ -4,6 +4,10 @@ metadata: name: {{ include "kafka-ui.fullname" . }} labels: {{- include "kafka-ui.labels" . | nindent 4 }} + {{- with .Values.annotations }} + annotations: + {{- toYaml . | nindent 4 }} + {{- end }} spec: {{- if not .Values.autoscaling.enabled }} replicas: {{ .Values.replicaCount }} diff --git a/charts/kafka-ui/values.yaml b/charts/kafka-ui/values.yaml index 93d08100d97..35e3ad3dd59 100644 --- a/charts/kafka-ui/values.yaml +++ b/charts/kafka-ui/values.yaml @@ -69,6 +69,10 @@ networkPolicy: podAnnotations: {} podLabels: {} +## Annotations to be added to kafka-ui Deployment +## +annotations: {} + podSecurityContext: {} # fsGroup: 2000
null
train
val
2022-11-03T18:12:59
"2022-10-14T08:58:33Z"
sherifkayad
train
provectus/kafka-ui/2378_2750
provectus/kafka-ui
provectus/kafka-ui/2378
provectus/kafka-ui/2750
[ "timestamp(timedelta=78.0, similarity=0.844880536159548)", "connected" ]
2b5dd270e22c6455c6a5999131dbf3ce3b770429
daba2a1f51686aa1260c9c9350883b1af79df04e
[ "#2632" ]
[]
"2022-10-14T09:48:55Z"
[ "type/enhancement", "good first issue", "status/accepted", "hacktoberfest", "scope/infrastructure", "scope/k8s" ]
Make it possible to update helm chart without an actual app release
[ ".github/workflows/helm.yaml" ]
[ ".github/workflows/helm.yaml" ]
[]
diff --git a/.github/workflows/helm.yaml b/.github/workflows/helm.yaml index 255f92260f9..d8c2875aed7 100644 --- a/.github/workflows/helm.yaml +++ b/.github/workflows/helm.yaml @@ -15,6 +15,18 @@ jobs: uses: Azure/setup-helm@v1 - name: Setup Kubeval uses: lra/[email protected] + #check, was helm version increased in Chart.yaml? + - name: Check version + shell: bash + run: | + git fetch + git checkout master + helm_version_old=$(cat charts/kafka-ui/Chart.yaml | grep version | awk '{print $2}') + git checkout $GITHUB_HEAD_REF + helm_version_new=$(cat charts/kafka-ui/Chart.yaml | grep version | awk '{print $2}') + echo $helm_version_old + echo $helm_version_new + if [[ "$helm_version_new" > "$helm_version_old" ]]; then exit 0 ; else exit 1 ; fi - name: Run kubeval shell: bash run: |
null
train
val
2022-10-14T11:16:29
"2022-08-02T13:28:33Z"
Haarolean
train
provectus/kafka-ui/2544_2756
provectus/kafka-ui
provectus/kafka-ui/2544
provectus/kafka-ui/2756
[ "keyword_pr_to_issue", "timestamp(timedelta=1.0, similarity=0.9183160250227483)" ]
b0782da7d137fba481eb9059a60e052617255b2a
fdf82986dc3adf262d52cc428a29cbc602aa1b7c
[]
[ "```suggestion\r\n let parsedJson;\r\n```", "```suggestion\r\n {item.field}: {get(parsedJson || {}, item.path)}\r\n```" ]
"2022-10-14T19:45:54Z"
[ "scope/frontend", "status/accepted", "type/feature" ]
Message fields preview
Add an ability to add message key/contents fields for preview (via json paths, probably) ![image](https://user-images.githubusercontent.com/1494347/189687995-6f148770-511c-4fc9-9b0e-a2b4273b0c47.png)
[ "kafka-ui-react-app/src/components/Topics/Topic/Messages/Message.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/__test__/Message.spec.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/__test__/MessagesTable.spec.tsx" ]
[ "kafka-ui-react-app/src/components/Topics/Topic/Messages/Message.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/PreviewModal.styled.ts", "kafka-ui-react-app/src/components/Topics/Topic/Messages/PreviewModal.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/__test__/Message.spec.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/__test__/MessagesTable.spec.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/__test__/PreviewModal.spec.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Message.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Message.tsx index fc4269bb32b..7efb1f728c1 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Message.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Message.tsx @@ -1,3 +1,4 @@ +import get from 'lodash/get'; import React from 'react'; import styled from 'styled-components'; import useDataSaver from 'lib/hooks/useDataSaver'; @@ -22,7 +23,14 @@ const ClickableRow = styled.tr` cursor: pointer; `; +export interface PreviewFilter { + field: string; + path: string; +} + export interface Props { + keyFilters: PreviewFilter[]; + contentFilters: PreviewFilter[]; message: TopicMessage; } @@ -38,6 +46,8 @@ const Message: React.FC<Props> = ({ keyFormat, headers, }, + keyFilters, + contentFilters, }) => { const [isOpen, setIsOpen] = React.useState(false); const savedMessageJson = { @@ -60,6 +70,33 @@ const Message: React.FC<Props> = ({ const [vEllipsisOpen, setVEllipsisOpen] = React.useState(false); + const getParsedJson = (jsonValue: string) => { + try { + return JSON.parse(jsonValue); + } catch (e) { + return {}; + } + }; + + const renderFilteredJson = ( + jsonValue?: string, + filters?: PreviewFilter[] + ) => { + if (!filters?.length || !jsonValue) return jsonValue; + + const parsedJson = getParsedJson(jsonValue); + + return ( + <> + {filters.map((item) => ( + <div> + {item.field}: {get(parsedJson, item.path)} + </div> + ))} + </> + ); + }; + return ( <> <ClickableRow @@ -77,10 +114,14 @@ const Message: React.FC<Props> = ({ <td> <div>{formatTimestamp(timestamp)}</div> </td> - <StyledDataCell title={key}>{key}</StyledDataCell> + <StyledDataCell title={key}> + {renderFilteredJson(key, keyFilters)} + </StyledDataCell> <StyledDataCell> <S.Metadata> - <S.MetadataValue>{content}</S.MetadataValue> + <S.MetadataValue> + {renderFilteredJson(content, contentFilters)} + </S.MetadataValue> </S.Metadata> </StyledDataCell> <td style={{ width: '5%' }}> diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx index 3deab1ecb5c..3c5901214b0 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx @@ -2,7 +2,7 @@ import PageLoader from 'components/common/PageLoader/PageLoader'; import { Table } from 'components/common/table/Table/Table.styled'; import TableHeaderCell from 'components/common/table/TableHeaderCell/TableHeaderCell'; import { TopicMessage } from 'generated-sources'; -import React, { useContext } from 'react'; +import React, { useContext, useState } from 'react'; import { getTopicMessges, getIsTopicMessagesFetching, @@ -10,14 +10,19 @@ import { import TopicMessagesContext from 'components/contexts/TopicMessagesContext'; import { useAppSelector } from 'lib/hooks/redux'; -import Message from './Message'; +import PreviewModal from './PreviewModal'; +import Message, { PreviewFilter } from './Message'; const MessagesTable: React.FC = () => { + const [previewFor, setPreviewFor] = useState<string | null>(null); + + const [keyFilters, setKeyFilters] = useState<PreviewFilter[]>([]); + const [contentFilters, setContentFilters] = useState<PreviewFilter[]>([]); + const { isLive } = useContext(TopicMessagesContext); const messages = useAppSelector(getTopicMessges); const isFetching = useAppSelector(getIsTopicMessagesFetching); - return ( <Table isFullwidth> <thead> @@ -26,9 +31,33 @@ const MessagesTable: React.FC = () => { <TableHeaderCell title="Offset" /> <TableHeaderCell title="Partition" /> <TableHeaderCell title="Timestamp" /> - <TableHeaderCell title="Key" /> - <TableHeaderCell title="Value" /> + <TableHeaderCell + title="Key" + previewText={`Preview ${ + keyFilters.length ? `(${keyFilters.length} selected)` : '' + }`} + onPreview={() => setPreviewFor('key')} + /> + <TableHeaderCell + title="Value" + previewText={`Preview ${ + contentFilters.length ? `(${contentFilters.length} selected)` : '' + }`} + onPreview={() => setPreviewFor('content')} + /> <TableHeaderCell> </TableHeaderCell> + + {previewFor !== null && ( + <PreviewModal + values={previewFor === 'key' ? keyFilters : contentFilters} + toggleIsOpen={() => setPreviewFor(null)} + setFilters={(payload: PreviewFilter[]) => + previewFor === 'key' + ? setKeyFilters(payload) + : setContentFilters(payload) + } + /> + )} </tr> </thead> <tbody> @@ -41,6 +70,8 @@ const MessagesTable: React.FC = () => { message.partition, ].join('-')} message={message} + keyFilters={keyFilters} + contentFilters={contentFilters} /> ))} {isFetching && isLive && !messages.length && ( diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/PreviewModal.styled.ts b/kafka-ui-react-app/src/components/Topics/Topic/Messages/PreviewModal.styled.ts new file mode 100644 index 00000000000..35f221ddeab --- /dev/null +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/PreviewModal.styled.ts @@ -0,0 +1,35 @@ +import styled from 'styled-components'; + +export const PreviewModal = styled.div` + height: auto; + width: 560px; + border-radius: 8px; + background: ${({ theme }) => theme.modal.backgroundColor}; + position: absolute; + left: 25%; + border: 1px solid ${({ theme }) => theme.modal.border.contrast}; + box-shadow: ${({ theme }) => theme.modal.shadow}; + padding: 32px; + z-index: 1; +`; + +export const ButtonWrapper = styled.div` + width: 100%; + display: flex; + justify-content: center; + margin-top: 20px; + gap: 10px; +`; + +export const EditForm = styled.div` + font-weight: 500; + padding-bottom: 7px; + display: flex; +`; + +export const Field = styled.div` + text-overflow: ellipsis; + white-space: nowrap; + overflow: hidden; + margin-right: 5px; +`; diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/PreviewModal.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Messages/PreviewModal.tsx new file mode 100644 index 00000000000..9d60f075d7d --- /dev/null +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/PreviewModal.tsx @@ -0,0 +1,137 @@ +import React, { useEffect } from 'react'; +import { Button } from 'components/common/Button/Button'; +import { FormError } from 'components/common/Input/Input.styled'; +import Input from 'components/common/Input/Input'; +import { InputLabel } from 'components/common/Input/InputLabel.styled'; +import IconButtonWrapper from 'components/common/Icons/IconButtonWrapper'; +import EditIcon from 'components/common/Icons/EditIcon'; +import CancelIcon from 'components/common/Icons/CancelIcon'; + +import * as S from './PreviewModal.styled'; +import { PreviewFilter } from './Message'; + +export interface InfoModalProps { + values: PreviewFilter[]; + toggleIsOpen(): void; + setFilters: (payload: PreviewFilter[]) => void; +} + +const PreviewModal: React.FC<InfoModalProps> = ({ + values, + toggleIsOpen, + setFilters, +}) => { + const [field, setField] = React.useState(''); + const [path, setPath] = React.useState(''); + const [errors, setErrors] = React.useState<string[]>([]); + const [editIndex, setEditIndex] = React.useState<number | undefined>(); + + const handleOk = () => { + const newErrors = []; + + if (field === '') { + newErrors.push('field'); + } + + if (path === '') { + newErrors.push('path'); + } + + if (newErrors?.length) { + setErrors(newErrors); + return; + } + + const newValues = [...values]; + + if (typeof editIndex !== 'undefined') { + newValues.splice(editIndex, 1, { field, path }); + } else { + newValues.push({ field, path }); + } + + setFilters(newValues); + toggleIsOpen(); + }; + + const handleRemove = (filter: PreviewFilter) => { + const newValues = values.filter( + (item) => item.field !== filter.field && item.path !== filter.path + ); + + setFilters(newValues); + }; + + useEffect(() => { + if (values?.length && typeof editIndex !== 'undefined') { + setField(values[editIndex].field); + setPath(values[editIndex].path); + } + }, [editIndex]); + + return ( + <S.PreviewModal> + {values.map((item, index) => ( + <S.EditForm key="index"> + <S.Field> + {' '} + {item.field} : {item.path} + </S.Field> + <IconButtonWrapper role="button" onClick={() => setEditIndex(index)}> + <EditIcon /> + </IconButtonWrapper> + {' '} + <IconButtonWrapper role="button" onClick={() => handleRemove(item)}> + <CancelIcon /> + </IconButtonWrapper> + </S.EditForm> + ))} + <div> + <InputLabel htmlFor="previewFormField">Field</InputLabel> + <Input + type="text" + id="previewFormField" + min="1" + value={field} + placeholder="Field" + onChange={({ target }) => setField(target?.value)} + /> + <FormError>{errors.includes('field') && 'Field is required'}</FormError> + </div> + <div> + <InputLabel htmlFor="previewFormJsonPath">Json path</InputLabel> + <Input + type="text" + id="previewFormJsonPath" + min="1" + value={path} + placeholder="Json Path" + onChange={({ target }) => setPath(target?.value)} + /> + <FormError> + {errors.includes('path') && 'Json path is required'} + </FormError> + </div> + <S.ButtonWrapper> + <Button + buttonSize="M" + buttonType="secondary" + type="button" + onClick={toggleIsOpen} + > + Close + </Button> + <Button + buttonSize="M" + buttonType="secondary" + type="button" + onClick={handleOk} + > + Save + </Button> + </S.ButtonWrapper> + </S.PreviewModal> + ); +}; + +export default PreviewModal; diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/__test__/Message.spec.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Messages/__test__/Message.spec.tsx index f70e2e0a42f..ec4d9cd8434 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/__test__/Message.spec.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/__test__/Message.spec.tsx @@ -1,6 +1,9 @@ import React from 'react'; import { TopicMessage, TopicMessageTimestampTypeEnum } from 'generated-sources'; -import Message, { Props } from 'components/Topics/Topic/Messages/Message'; +import Message, { + PreviewFilter, + Props, +} from 'components/Topics/Topic/Messages/Message'; import { screen } from '@testing-library/react'; import { render } from 'lib/testHelpers'; import userEvent from '@testing-library/user-event'; @@ -30,6 +33,9 @@ describe('Message component', () => { headers: { header: 'test' }, }; + const mockKeyFilters: PreviewFilter[] = []; + const mockContentFilters: PreviewFilter[] = []; + const renderComponent = ( props: Partial<Props> = { message: mockMessage, @@ -38,7 +44,11 @@ describe('Message component', () => { return render( <table> <tbody> - <Message message={props.message || mockMessage} /> + <Message + message={props.message || mockMessage} + keyFilters={mockKeyFilters} + contentFilters={mockContentFilters} + /> </tbody> </table> ); diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/__test__/MessagesTable.spec.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Messages/__test__/MessagesTable.spec.tsx index fa78e364a73..37ad28b202b 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/__test__/MessagesTable.spec.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/__test__/MessagesTable.spec.tsx @@ -1,5 +1,6 @@ import React from 'react'; import { screen } from '@testing-library/react'; +import userEvent from '@testing-library/user-event'; import { render } from 'lib/testHelpers'; import MessagesTable from 'components/Topics/Topic/Messages/MessagesTable'; import { SeekDirection, SeekType, TopicMessage } from 'generated-sources'; @@ -65,6 +66,19 @@ describe('MessagesTable', () => { expect(screen.getByRole('table')).toBeInTheDocument(); }); + it('should check preview buttons', async () => { + const previewButtons = await screen.findAllByRole('button', { + name: 'Preview', + }); + expect(previewButtons).toHaveLength(2); + }); + + it('should show preview modal with validation', async () => { + await userEvent.click(screen.getAllByText('Preview')[0]); + expect(screen.getByPlaceholderText('Field')).toHaveValue(''); + expect(screen.getByPlaceholderText('Json Path')).toHaveValue(''); + }); + it('should check the if no elements is rendered in the table', () => { expect(screen.getByText(/No messages found/i)).toBeInTheDocument(); }); diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/__test__/PreviewModal.spec.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Messages/__test__/PreviewModal.spec.tsx new file mode 100644 index 00000000000..fbea81562ee --- /dev/null +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/__test__/PreviewModal.spec.tsx @@ -0,0 +1,112 @@ +import userEvent from '@testing-library/user-event'; +import { act, screen } from '@testing-library/react'; +import { render } from 'lib/testHelpers'; +import React from 'react'; +import { PreviewFilter } from 'components/Topics/Topic/Messages/Message'; +import { serdesPayload } from 'lib/fixtures/topicMessages'; +import { useSerdes } from 'lib/hooks/api/topicMessages'; +import PreviewModal, { + InfoModalProps, +} from 'components/Topics/Topic/Messages/PreviewModal'; + +jest.mock('components/common/Icons/CloseIcon', () => () => 'mock-CloseIcon'); + +jest.mock('lib/hooks/api/topicMessages', () => ({ + useSerdes: jest.fn(), +})); + +beforeEach(async () => { + (useSerdes as jest.Mock).mockImplementation(() => ({ + data: serdesPayload, + })); +}); + +const toggleInfoModal = jest.fn(); +const mockValues: PreviewFilter[] = [ + { + field: '', + path: '', + }, +]; + +const renderComponent = (props?: Partial<InfoModalProps>) => { + render( + <PreviewModal + toggleIsOpen={toggleInfoModal} + values={mockValues} + setFilters={jest.fn()} + {...props} + /> + ); +}; + +describe('PreviewModal component', () => { + it('closes PreviewModal', async () => { + renderComponent(); + await userEvent.click(screen.getByRole('button', { name: 'Close' })); + expect(toggleInfoModal).toHaveBeenCalledTimes(1); + }); + + it('return if empty inputs', async () => { + renderComponent(); + await userEvent.click(screen.getByRole('button', { name: 'Save' })); + expect(screen.getByText('Json path is required')).toBeInTheDocument(); + expect(screen.getByText('Field is required')).toBeInTheDocument(); + }); + + describe('Input elements', () => { + const fieldValue = 'type'; + const pathValue = 'schema.type'; + + beforeEach(async () => { + await act(() => { + renderComponent(); + }); + }); + + it('field input', async () => { + const fieldInput = screen.getByPlaceholderText('Field'); + expect(fieldInput).toHaveValue(''); + await userEvent.type(fieldInput, fieldValue); + expect(fieldInput).toHaveValue(fieldValue); + }); + + it('path input', async () => { + const pathInput = screen.getByPlaceholderText('Json Path'); + expect(pathInput).toHaveValue(''); + await userEvent.type(pathInput, pathValue); + expect(pathInput).toHaveValue(pathValue.toString()); + }); + }); + + describe('edit and remove functionality', () => { + const fieldValue = 'type new'; + const pathValue = 'schema.type.new'; + + it('remove values', async () => { + const setFilters = jest.fn(); + await act(() => { + renderComponent({ setFilters }); + }); + await userEvent.click(screen.getByRole('button', { name: 'Cancel' })); + expect(setFilters).toHaveBeenCalledTimes(1); + }); + + it('edit values', async () => { + const setFilters = jest.fn(); + const toggleIsOpen = jest.fn(); + await act(() => { + renderComponent({ setFilters }); + }); + userEvent.click(screen.getByRole('button', { name: 'Edit' })); + const fieldInput = screen.getByPlaceholderText('Field'); + userEvent.type(fieldInput, fieldValue); + const pathInput = screen.getByPlaceholderText('Json Path'); + userEvent.type(pathInput, pathValue); + userEvent.click(screen.getByRole('button', { name: 'Save' })); + await act(() => { + renderComponent({ setFilters, toggleIsOpen }); + }); + }); + }); +});
null
train
val
2022-11-07T14:01:57
"2022-09-08T03:43:04Z"
armenuikafka
train
provectus/kafka-ui/2675_2765
provectus/kafka-ui
provectus/kafka-ui/2675
provectus/kafka-ui/2765
[ "keyword_pr_to_issue" ]
63b6bb3fa10dd33b069c15151bf706a313970550
0e4a314082d6e5d64ba0d09e60ebff8f5170aac7
[ "Ready to take this issue.", "So I started this, but not able to build the UI locally. I followed the instructions here:\r\n\r\nhttps://github.com/provectus/kafka-ui/tree/master/kafka-ui-react-app#getting-started\r\n\r\nMy node_modules get installed and I launch the docker containers successfully. The command to start runs as well, and when I go to localhost:3000, I do not see the UI built after I made the changes.\r\n\r\nAm I doing something wrong?", "Oof, what a competition, folks. I'm not sure who should I assign this to -- the first person who asked or the one who decided to start anyway? Please check out our \"up for grabs\" board/project for other issues. We don't have much frontend ones right now, but there are still some left.", "> So I started this, but not able to build the UI locally. I followed the instructions here:\r\n> \r\n> https://github.com/provectus/kafka-ui/tree/master/kafka-ui-react-app#getting-started\r\n> \r\n> My node_modules get installed and I launch the docker containers successfully. The command to start runs as well, and when I go to localhost:3000, I do not see the UI built after I made the changes.\r\n> \r\n> Am I doing something wrong?\r\n\r\nWhat do you observe there? Do you mean your changes are not reflected there? AFAIK vite does the rebuild in runtime just fine.", "Anyone's willing to work on this? ", "Me🀚", "Do I also need to change MessagesV2 ?", "Yes, please :)", "Hey, I am a little confused...\r\nAre this the following files in kafki-ui-react-app where change is to be made?\r\n`src/components/Topics/Topic/Messages/Message.tsx`\r\n`src/components/Topics/Topic/Messages/MessagesTable.tsx`\r\n`src/components/Topics/Topic/MessagesV2/ActionCell.tsx`\r\n`src/components/Topics/Topic/MessagesV2/MessagesTable.tsx`\r\n\r\nAnd sorry for being late.", "@adityabhattad2021 if I'll start looking for all the files, I'd do the same work for the second time :)\r\n\r\nThese are the places on sshots, and there might be others.\r\n\r\n![Untitled](https://user-images.githubusercontent.com/1494347/196151351-480c4a44-eac9-48fc-be80-73e68bab47d1.jpg)\r\n![Untitled2](https://user-images.githubusercontent.com/1494347/196151412-b22a51cc-fc26-43bf-a418-ae5b060cade4.jpg)\r\n\r\n", "Could we please update also in message preview ? with expanding the message \r\n\r\n<img width=\"1725\" alt=\"value\" src=\"https://user-images.githubusercontent.com/104780608/196926364-fea26357-d826-47c4-9e72-92de597f12d1.png\">\r\n", "I'll do it.", "Hey I was not able to run react app locally by following the documentation for Docker way\r\nThis was the error\r\n![Screenshot 2022-10-24 180823](https://user-images.githubusercontent.com/93488388/197530668-d7904689-11aa-4891-9c27-08c6ac4dd325.png)\r\n", "> Hey I was not able to run react app locally by following the documentation for Docker way This was the error ![Screenshot 2022-10-24 180823](https://user-images.githubusercontent.com/93488388/197530668-d7904689-11aa-4891-9c27-08c6ac4dd325.png)\r\n\r\nit seems like you don't have the docker installed?", "I did installed and setup docker.", "> I did installed and setup docker.\r\n\r\nI don't see the initial command which caused this error. ", "I have found one more content label instead of value when downloading the Kafka msg. Is this intentional (not to break JSON parsers) or is it a bug?\r\n\r\n```json\r\n{\r\n\t\"Content\": \"Lorem ipsum dolor sit amet\",\r\n\t\"Offset\": 296897,\r\n\t\"Key\": \"12345\",\r\n\t\"Partition\": 0,\r\n\t\"Headers\": {},\r\n\t\"Timestamp\": \"2023-01-17T12:00:02.437Z\"\r\n}\r\n```", "@DanielSmetanka probably a bug, I'll raise an issue" ]
[]
"2022-10-17T13:17:05Z"
[ "good first issue", "scope/frontend", "status/accepted", "type/chore", "hacktoberfest" ]
Topic messages: rename "content" to "value"
Everywhere in the messages views rename "content" to "value". It's key/value in kafka, not key/content.
[ "kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx", "kafka-ui-react-app/src/components/Topics/Topic/SendMessage/SendMessage.tsx" ]
[ "kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx", "kafka-ui-react-app/src/components/Topics/Topic/SendMessage/SendMessage.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx index 3ee4b552efa..c010ca65cb5 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx @@ -459,7 +459,7 @@ const Filters: React.FC<FiltersProps> = ({ /> </div> <div> - <InputLabel>Content Serde</InputLabel> + <InputLabel>Value Serde</InputLabel> <Select id="selectValueSerdeOptions" aria-labelledby="selectValueSerdeOptions" diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx index 4696b75e036..3deab1ecb5c 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx @@ -27,7 +27,7 @@ const MessagesTable: React.FC = () => { <TableHeaderCell title="Partition" /> <TableHeaderCell title="Timestamp" /> <TableHeaderCell title="Key" /> - <TableHeaderCell title="Content" /> + <TableHeaderCell title="Value" /> <TableHeaderCell> </TableHeaderCell> </tr> </thead> diff --git a/kafka-ui-react-app/src/components/Topics/Topic/SendMessage/SendMessage.tsx b/kafka-ui-react-app/src/components/Topics/Topic/SendMessage/SendMessage.tsx index 087abe537ae..b0c35fc7504 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/SendMessage/SendMessage.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/SendMessage/SendMessage.tsx @@ -154,7 +154,7 @@ const SendMessage: React.FC<{ onSubmit: () => void }> = ({ onSubmit }) => { /> </S.Column> <S.Column> - <InputLabel>Content Serde</InputLabel> + <InputLabel>Value Serde</InputLabel> <Controller control={control} name="valueSerde" @@ -190,7 +190,7 @@ const SendMessage: React.FC<{ onSubmit: () => void }> = ({ onSubmit }) => { /> </S.Column> <S.Column> - <InputLabel>Content</InputLabel> + <InputLabel>Value</InputLabel> <Controller control={control} name="content"
null
train
val
2022-10-17T14:24:14
"2022-09-30T09:19:57Z"
Haarolean
train
provectus/kafka-ui/2705_2766
provectus/kafka-ui
provectus/kafka-ui/2705
provectus/kafka-ui/2766
[ "connected", "timestamp(timedelta=0.0, similarity=0.9654116107887731)" ]
dabe2878c1610fc6631d3ab81bb7aa7048617a47
1b8ee3b6723d2159217c39a43166e96035c4c311
[]
[ "what if there is no page property?" ]
"2022-10-17T14:55:49Z"
[ "type/bug", "good first issue", "scope/frontend", "status/accepted", "status/confirmed", "hacktoberfest" ]
[UI] Topic search not working properly with pagination
**Setup:** version in c961344 **Description:** 1. Go to Topics page 2. move to page 2 3. enter something into topic search field -> request sent with `page=2` parameter which will force BE to search **on the second page**, which will lead to wrong results <img width="904" alt="Screenshot 2022-10-06 at 10 59 46" src="https://user-images.githubusercontent.com/702205/194235461-dbd248e0-ef72-4079-9b6f-96ae37d54224.png"> **Expected behaviour:** When something new was entered into search field pagination should be reset --- **Note:** such behaviour should be fixed in all places where pagination is used: SchemaRegistry, Consumers
[ "kafka-ui-react-app/src/components/common/Search/Search.tsx" ]
[ "kafka-ui-react-app/src/components/common/Search/Search.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/common/Search/Search.tsx b/kafka-ui-react-app/src/components/common/Search/Search.tsx index 9baf1ed3b2e..66c0e95030b 100644 --- a/kafka-ui-react-app/src/components/common/Search/Search.tsx +++ b/kafka-ui-react-app/src/components/common/Search/Search.tsx @@ -22,6 +22,9 @@ const Search: React.FC<SearchProps> = ({ onChange(e.target.value); } else { searchParams.set('q', e.target.value); + if (searchParams.get('page')) { + searchParams.set('page', '1'); + } setSearchParams(searchParams); } }, 500);
null
train
val
2022-10-24T21:18:18
"2022-10-06T07:02:14Z"
iliax
train
provectus/kafka-ui/2680_2771
provectus/kafka-ui
provectus/kafka-ui/2680
provectus/kafka-ui/2771
[ "connected" ]
aa6c3083c2c44aa2a1764ea286d5fca9a34b4f9c
2c74b4e1c2ddedcd0f73843eb630ed98340cf3dd
[]
[]
"2022-10-18T07:03:36Z"
[ "type/enhancement", "scope/backend", "scope/frontend", "status/accepted" ]
UI: Display smart filters filtered stats
When polling messages from topic (and filtering is enabled) we should show number of errors that appeared during filtration process - > this number placed in `TopicMessageConsuming.filterApplyErrors` UI: 1. during polling process - show value of `TopicMessageConsuming.filterApplyErrors` if it is > 0 2. when polling is finished BE will send `TopicMessageEvent` with `type = DONE` with final stats of polling process - this stats should be rendered. (also this event should be logically treated as polling end by UI)
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/AbstractEmitter.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/BackwardRecordEmitter.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/ConsumingStats.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/ForwardRecordEmitter.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/MessageFilters.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/AdminClientServiceImpl.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/MessagesService.java", "kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml" ]
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/AbstractEmitter.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/BackwardRecordEmitter.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/ConsumingStats.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/ForwardRecordEmitter.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/MessageFilterStats.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/MessageFilters.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/AdminClientServiceImpl.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/MessagesService.java", "kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml" ]
[ "kafka-ui-api/src/test/java/com/provectus/kafka/ui/service/RecordEmitterTest.java" ]
diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/AbstractEmitter.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/AbstractEmitter.java index ddb5dc3666c..35d8a3bef34 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/AbstractEmitter.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/AbstractEmitter.java @@ -64,6 +64,18 @@ protected void sendPhase(FluxSink<TopicMessageEventDTO> sink, String name) { protected void sendConsuming(FluxSink<TopicMessageEventDTO> sink, ConsumerRecords<Bytes, Bytes> records, long elapsed) { - consumingStats.sendConsumingEvt(sink, records, elapsed); + consumingStats.sendConsumingEvt(sink, records, elapsed, getFilterApplyErrors(sink)); + } + + protected void sendFinishStatsAndCompleteSink(FluxSink<TopicMessageEventDTO> sink) { + consumingStats.sendFinishEvent(sink, getFilterApplyErrors(sink)); + sink.complete(); + } + + protected Number getFilterApplyErrors(FluxSink<?> sink) { + return sink.contextView() + .<MessageFilterStats>getOrEmpty(MessageFilterStats.class) + .<Number>map(MessageFilterStats::getFilterApplyErrors) + .orElse(0); } } \ No newline at end of file diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/BackwardRecordEmitter.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/BackwardRecordEmitter.java index d2012355db6..fe1644676a0 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/BackwardRecordEmitter.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/BackwardRecordEmitter.java @@ -80,7 +80,7 @@ public void accept(FluxSink<TopicMessageEventDTO> sink) { log.debug("sink is cancelled after partitions poll iteration"); } } - sink.complete(); + sendFinishStatsAndCompleteSink(sink); log.debug("Polling finished"); } catch (Exception e) { log.error("Error occurred while consuming records", e); diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/ConsumingStats.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/ConsumingStats.java index 830eb87320b..bd0754fc687 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/ConsumingStats.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/ConsumingStats.java @@ -16,7 +16,8 @@ class ConsumingStats { void sendConsumingEvt(FluxSink<TopicMessageEventDTO> sink, ConsumerRecords<Bytes, Bytes> polledRecords, - long elapsed) { + long elapsed, + Number filterApplyErrors) { for (ConsumerRecord<Bytes, Bytes> rec : polledRecords) { for (Header header : rec.headers()) { bytes += @@ -27,15 +28,28 @@ void sendConsumingEvt(FluxSink<TopicMessageEventDTO> sink, } this.records += polledRecords.count(); this.elapsed += elapsed; - final TopicMessageConsumingDTO consuming = new TopicMessageConsumingDTO() - .bytesConsumed(this.bytes) - .elapsedMs(this.elapsed) - .isCancelled(sink.isCancelled()) - .messagesConsumed(this.records); sink.next( new TopicMessageEventDTO() .type(TopicMessageEventDTO.TypeEnum.CONSUMING) - .consuming(consuming) + .consuming(createConsumingStats(sink, filterApplyErrors)) ); } + + void sendFinishEvent(FluxSink<TopicMessageEventDTO> sink, Number filterApplyErrors) { + sink.next( + new TopicMessageEventDTO() + .type(TopicMessageEventDTO.TypeEnum.DONE) + .consuming(createConsumingStats(sink, filterApplyErrors)) + ); + } + + private TopicMessageConsumingDTO createConsumingStats(FluxSink<TopicMessageEventDTO> sink, + Number filterApplyErrors) { + return new TopicMessageConsumingDTO() + .bytesConsumed(this.bytes) + .elapsedMs(this.elapsed) + .isCancelled(sink.isCancelled()) + .filterApplyErrors(filterApplyErrors.intValue()) + .messagesConsumed(this.records); + } } diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/ForwardRecordEmitter.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/ForwardRecordEmitter.java index 69d9801b705..502a8c9a26c 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/ForwardRecordEmitter.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/ForwardRecordEmitter.java @@ -54,7 +54,7 @@ public void accept(FluxSink<TopicMessageEventDTO> sink) { } } } - sink.complete(); + sendFinishStatsAndCompleteSink(sink); log.info("Polling finished"); } catch (Exception e) { log.error("Error occurred while consuming records", e); diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/MessageFilterStats.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/MessageFilterStats.java new file mode 100644 index 00000000000..3b6df3cdea4 --- /dev/null +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/MessageFilterStats.java @@ -0,0 +1,16 @@ +package com.provectus.kafka.ui.emitter; + +import java.util.concurrent.atomic.AtomicLong; +import lombok.AccessLevel; +import lombok.Getter; + +public class MessageFilterStats { + + @Getter(AccessLevel.PACKAGE) + private final AtomicLong filterApplyErrors = new AtomicLong(); + + public final void incrementApplyErrors() { + filterApplyErrors.incrementAndGet(); + } + +} diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/MessageFilters.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/MessageFilters.java index ea0707f1029..e48501f6a75 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/MessageFilters.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/MessageFilters.java @@ -9,6 +9,7 @@ import javax.script.CompiledScript; import javax.script.ScriptEngineManager; import javax.script.ScriptException; +import lombok.SneakyThrows; import lombok.extern.slf4j.Slf4j; import org.apache.commons.lang3.StringUtils; import org.codehaus.groovy.jsr223.GroovyScriptEngineImpl; @@ -40,25 +41,26 @@ static Predicate<TopicMessageDTO> containsStringFilter(String string) { static Predicate<TopicMessageDTO> groovyScriptFilter(String script) { var compiledScript = compileScript(script); var jsonSlurper = new JsonSlurper(); - return msg -> { - var bindings = getGroovyEngine().createBindings(); - bindings.put("partition", msg.getPartition()); - bindings.put("offset", msg.getOffset()); - bindings.put("timestampMs", msg.getTimestamp().toInstant().toEpochMilli()); - bindings.put("keyAsText", msg.getKey()); - bindings.put("valueAsText", msg.getContent()); - bindings.put("headers", msg.getHeaders()); - bindings.put("key", parseToJsonOrReturnNull(jsonSlurper, msg.getKey())); - bindings.put("value", parseToJsonOrReturnNull(jsonSlurper, msg.getContent())); - try { + return new Predicate<TopicMessageDTO>() { + @SneakyThrows + @Override + public boolean test(TopicMessageDTO msg) { + var bindings = getGroovyEngine().createBindings(); + bindings.put("partition", msg.getPartition()); + bindings.put("offset", msg.getOffset()); + bindings.put("timestampMs", msg.getTimestamp().toInstant().toEpochMilli()); + bindings.put("keyAsText", msg.getKey()); + bindings.put("valueAsText", msg.getContent()); + bindings.put("headers", msg.getHeaders()); + bindings.put("key", parseToJsonOrReturnNull(jsonSlurper, msg.getKey())); + bindings.put("value", parseToJsonOrReturnNull(jsonSlurper, msg.getContent())); var result = compiledScript.eval(bindings); if (result instanceof Boolean) { return (Boolean) result; + } else { + throw new ValidationException( + String.format("Unexpected script result: %s, Boolean should be returned instead", result)); } - return false; - } catch (Exception e) { - log.trace("Error executing filter script '{}' on message '{}' ", script, msg, e); - return false; } }; } diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/AdminClientServiceImpl.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/AdminClientServiceImpl.java index 7b93b1420e0..9291198df73 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/AdminClientServiceImpl.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/AdminClientServiceImpl.java @@ -37,6 +37,7 @@ private Mono<ReactiveAdminClient> createAdminClient(KafkaCluster cluster) { properties .put(AdminClientConfig.BOOTSTRAP_SERVERS_CONFIG, cluster.getBootstrapServers()); properties.put(AdminClientConfig.REQUEST_TIMEOUT_MS_CONFIG, clientTimeout); + properties.putIfAbsent(AdminClientConfig.CLIENT_ID_CONFIG, "kafka-ui-app"); return AdminClient.create(properties); }) .flatMap(ReactiveAdminClient::create) diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/MessagesService.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/MessagesService.java index ecfeda01227..d0bf94c98b6 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/MessagesService.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/MessagesService.java @@ -2,6 +2,7 @@ import com.provectus.kafka.ui.emitter.BackwardRecordEmitter; import com.provectus.kafka.ui.emitter.ForwardRecordEmitter; +import com.provectus.kafka.ui.emitter.MessageFilterStats; import com.provectus.kafka.ui.emitter.MessageFilters; import com.provectus.kafka.ui.emitter.TailingEmitter; import com.provectus.kafka.ui.exception.TopicNotFoundException; @@ -172,8 +173,10 @@ private Flux<TopicMessageEventDTO> loadMessagesImpl(KafkaCluster cluster, recordDeserializer ); } + MessageFilterStats filterStats = new MessageFilterStats(); return Flux.create(emitter) - .filter(getMsgFilter(query, filterQueryType)) + .contextWrite(ctx -> ctx.put(MessageFilterStats.class, filterStats)) + .filter(getMsgFilter(query, filterQueryType, filterStats)) .takeWhile(createTakeWhilePredicate(seekDirection, limit)) .subscribeOn(Schedulers.boundedElastic()) .share(); @@ -186,7 +189,9 @@ private Predicate<TopicMessageEventDTO> createTakeWhilePredicate( : new ResultSizeLimiter(limit); } - private Predicate<TopicMessageEventDTO> getMsgFilter(String query, MessageFilterTypeDTO filterQueryType) { + private Predicate<TopicMessageEventDTO> getMsgFilter(String query, + MessageFilterTypeDTO filterQueryType, + MessageFilterStats filterStats) { if (StringUtils.isEmpty(query)) { return evt -> true; } @@ -194,7 +199,13 @@ private Predicate<TopicMessageEventDTO> getMsgFilter(String query, MessageFilter return evt -> { // we only apply filter for message events if (evt.getType() == TopicMessageEventDTO.TypeEnum.MESSAGE) { - return messageFilter.test(evt.getMessage()); + try { + return messageFilter.test(evt.getMessage()); + } catch (Exception e) { + filterStats.incrementApplyErrors(); + log.trace("Error applying filter '{}' for message {}", query, evt.getMessage()); + return false; + } } return true; }; diff --git a/kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml b/kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml index a1ba87d85af..1034fc32025 100644 --- a/kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml +++ b/kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml @@ -2408,6 +2408,8 @@ components: type: boolean messagesConsumed: type: integer + filterApplyErrors: + type: integer TopicMessage:
diff --git a/kafka-ui-api/src/test/java/com/provectus/kafka/ui/service/RecordEmitterTest.java b/kafka-ui-api/src/test/java/com/provectus/kafka/ui/service/RecordEmitterTest.java index 3289d177d2b..350386fd7e9 100644 --- a/kafka-ui-api/src/test/java/com/provectus/kafka/ui/service/RecordEmitterTest.java +++ b/kafka-ui-api/src/test/java/com/provectus/kafka/ui/service/RecordEmitterTest.java @@ -120,17 +120,17 @@ void pollNothingOnEmptyTopic() { RECORD_DESERIALIZER ); - StepVerifier.create( - Flux.create(forwardEmitter) - .filter(m -> m.getType().equals(TopicMessageEventDTO.TypeEnum.MESSAGE)) - .take(100) - ).expectNextCount(0).expectComplete().verify(); - - StepVerifier.create( - Flux.create(backwardEmitter) - .filter(m -> m.getType().equals(TopicMessageEventDTO.TypeEnum.MESSAGE)) - .take(100) - ).expectNextCount(0).expectComplete().verify(); + StepVerifier.create(Flux.create(forwardEmitter)) + .expectNextMatches(m -> m.getType().equals(TopicMessageEventDTO.TypeEnum.PHASE)) + .expectNextMatches(m -> m.getType().equals(TopicMessageEventDTO.TypeEnum.DONE)) + .expectComplete() + .verify(); + + StepVerifier.create(Flux.create(backwardEmitter)) + .expectNextMatches(m -> m.getType().equals(TopicMessageEventDTO.TypeEnum.PHASE)) + .expectNextMatches(m -> m.getType().equals(TopicMessageEventDTO.TypeEnum.DONE)) + .expectComplete() + .verify(); } @Test
train
val
2022-10-31T13:46:19
"2022-09-30T09:33:59Z"
Haarolean
train
provectus/kafka-ui/2679_2775
provectus/kafka-ui
provectus/kafka-ui/2679
provectus/kafka-ui/2775
[ "timestamp(timedelta=1.0, similarity=0.959965375729226)", "connected" ]
ee920b43c45156cad1b89be31e43106ca3859edb
b46ee9c5ccd2e8df436e05e76fad393bdd4c5179
[]
[ "create separate object in root \r\n```\r\ntooltip: {\r\n bg:\r\n text:\r\n}\r\n```" ]
"2022-10-18T13:30:17Z"
[ "good first issue", "scope/frontend", "status/accepted", "type/chore" ]
Broker configs: Display a tooltip about the source of the property
<img width="257" alt="image" src="https://user-images.githubusercontent.com/1494347/193239459-4583ebbf-fde1-417c-b4e5-a0a9d2fb5cf8.png"> Add a tooltip icon with the following text on hover: ``` DYNAMIC_TOPIC_CONFIG: dynamic topic config that is configured for a specific topic DYNAMIC_BROKER_LOGGER_CONFIG: dynamic broker logger config that is configured for a specific broker DYNAMIC_BROKER_CONFIG: dynamic broker config that is configured for a specific broker DYNAMIC_DEFAULT_BROKER_CONFIG: dynamic broker config that is configured as default for all brokers in the cluster STATIC_BROKER_CONFIG: static broker config provided as broker properties at start up (e.g. server.properties file) DEFAULT_CONFIG: built-in default configuration for configs that have a default value UNKNOWN: source unknown e.g. in the ConfigEntry used for alter requests where source is not set ```
[ "kafka-ui-react-app/package.json", "kafka-ui-react-app/pnpm-lock.yaml", "kafka-ui-react-app/src/components/Brokers/Broker/Configs/Configs.styled.ts", "kafka-ui-react-app/src/components/Brokers/Broker/Configs/Configs.tsx", "kafka-ui-react-app/src/theme/theme.ts" ]
[ "kafka-ui-react-app/package.json", "kafka-ui-react-app/pnpm-lock.yaml", "kafka-ui-react-app/src/components/Brokers/Broker/Configs/Configs.styled.ts", "kafka-ui-react-app/src/components/Brokers/Broker/Configs/Configs.tsx", "kafka-ui-react-app/src/components/common/Icons/InfoIcon.tsx", "kafka-ui-react-app/src/components/common/Tooltip/Tooltip.styled.ts", "kafka-ui-react-app/src/components/common/Tooltip/Tooltip.tsx", "kafka-ui-react-app/src/components/common/Tooltip/__tests__/Tooltip.spec.tsx", "kafka-ui-react-app/src/theme/theme.ts" ]
[]
diff --git a/kafka-ui-react-app/package.json b/kafka-ui-react-app/package.json index fa1842dedbd..b0779e54884 100644 --- a/kafka-ui-react-app/package.json +++ b/kafka-ui-react-app/package.json @@ -7,6 +7,7 @@ "@babel/core": "^7.16.0", "@babel/plugin-syntax-flow": "^7.18.6", "@babel/plugin-transform-react-jsx": "^7.18.6", + "@floating-ui/react-dom-interactions": "^0.10.3", "@hookform/error-message": "^2.0.0", "@hookform/resolvers": "^2.7.1", "@microsoft/fetch-event-source": "^2.0.1", diff --git a/kafka-ui-react-app/pnpm-lock.yaml b/kafka-ui-react-app/pnpm-lock.yaml index 88fcfac42e1..dca028c2d9f 100644 --- a/kafka-ui-react-app/pnpm-lock.yaml +++ b/kafka-ui-react-app/pnpm-lock.yaml @@ -7,6 +7,7 @@ specifiers: '@babel/preset-env': ^7.18.2 '@babel/preset-react': ^7.17.12 '@babel/preset-typescript': ^7.17.12 + '@floating-ui/react-dom-interactions': ^0.10.3 '@hookform/error-message': ^2.0.0 '@hookform/resolvers': ^2.7.1 '@jest/types': ^29.0.3 @@ -94,6 +95,7 @@ dependencies: '@babel/core': 7.18.2 '@babel/plugin-syntax-flow': 7.18.6_@[email protected] '@babel/plugin-transform-react-jsx': 7.18.6_@[email protected] + '@floating-ui/react-dom-interactions': 0.10.3_ohobp6rpsmerwlq5ipwfh5yigy '@hookform/error-message': 2.0.0_l2dcsysovzdujulgxvsen7vbsm '@hookform/resolvers': [email protected] '@microsoft/fetch-event-source': 2.0.1 @@ -2733,6 +2735,41 @@ packages: - supports-color dev: true + /@floating-ui/core/1.0.1: + resolution: {integrity: sha512-bO37brCPfteXQfFY0DyNDGB3+IMe4j150KFQcgJ5aBP295p9nBGeHEs/p0czrRbtlHq4Px/yoPXO/+dOCcF4uA==} + dev: false + + /@floating-ui/dom/1.0.4: + resolution: {integrity: sha512-maYJRv+sAXTy4K9mzdv0JPyNW5YPVHrqtY90tEdI6XNpuLOP26Ci2pfwPsKBA/Wh4Z3FX5sUrtUFTdMYj9v+ug==} + dependencies: + '@floating-ui/core': 1.0.1 + dev: false + + /@floating-ui/react-dom-interactions/0.10.3_ohobp6rpsmerwlq5ipwfh5yigy: + resolution: {integrity: sha512-UEHqdnzyoiWNU5az/tAljr9iXFzN18DcvpMqW+/cXz4FEhDEB1ogLtWldOWCujLerPBnSRocADALafelOReMpw==} + peerDependencies: + react: '>=16.8.0' + react-dom: '>=16.8.0' + dependencies: + '@floating-ui/react-dom': 1.0.0_ef5jwxihqo6n7gxfmzogljlgcm + aria-hidden: 1.2.1_7cpxmzzodpxnolj5zcc5cr63ji + react: 18.1.0 + react-dom: [email protected] + transitivePeerDependencies: + - '@types/react' + dev: false + + /@floating-ui/react-dom/1.0.0_ef5jwxihqo6n7gxfmzogljlgcm: + resolution: {integrity: sha512-uiOalFKPG937UCLm42RxjESTWUVpbbatvlphQAU6bsv+ence6IoVG8JOUZcy8eW81NkU+Idiwvx10WFLmR4MIg==} + peerDependencies: + react: '>=16.8.0' + react-dom: '>=16.8.0' + dependencies: + '@floating-ui/dom': 1.0.4 + react: 18.1.0 + react-dom: [email protected] + dev: false + /@hookform/error-message/2.0.0_l2dcsysovzdujulgxvsen7vbsm: resolution: {integrity: sha512-Y90nHzjgL2MP7GFy75kscdvxrCTjtyxGmOLLxX14nd08OXRIh9lMH/y9Kpdo0p1IPowJBiZMHyueg7p+yrqynQ==} peerDependencies: @@ -3899,6 +3936,21 @@ packages: resolution: {integrity: sha512-8+9WqebbFzpX9OR+Wa6O29asIogeRMzcGtAINdpMHHyAg10f05aSFVBbcEqGf/PXw1EjAZ+q2/bEBg3DvurK3Q==} dev: true + /aria-hidden/1.2.1_7cpxmzzodpxnolj5zcc5cr63ji: + resolution: {integrity: sha512-PN344VAf9j1EAi+jyVHOJ8XidQdPVssGco39eNcsGdM4wcsILtxrKLkbuiMfLWYROK1FjRQasMWCBttrhjnr6A==} + engines: {node: '>=10'} + peerDependencies: + '@types/react': ^16.9.0 || ^17.0.0 || ^18.0.0 + react: ^16.9.0 || ^17.0.0 || ^18.0.0 + peerDependenciesMeta: + '@types/react': + optional: true + dependencies: + '@types/react': 18.0.9 + react: 18.1.0 + tslib: 2.4.0 + dev: false + /aria-query/4.2.2: resolution: {integrity: sha512-o/HelwhuKpTj/frsOsbNLNgnNGVIFsVP/SW2BSF14gVl7kAfMOJ6/8wUAUvG1R1NHKrfG+2sHZTu0yauT1qBrA==} engines: {node: '>=6.0'} @@ -8531,7 +8583,6 @@ packages: /tslib/2.4.0: resolution: {integrity: sha512-d6xOpEDfsi2CZVlPQzGeux8XMwLT9hssAsaPYExaQMuYskwb+x1x7J371tWlbBdWHroy99KnVB6qIkUbs5X3UQ==} - dev: true /tsutils/[email protected]: resolution: {integrity: sha512-mHKK3iUXL+3UF6xL5k0PEhKRUBKPBCv/+RkEOpjRWxxx27KKRBmmA60A9pgOUvMi8GKhRMPEmjBRPzs2W7O1OA==} diff --git a/kafka-ui-react-app/src/components/Brokers/Broker/Configs/Configs.styled.ts b/kafka-ui-react-app/src/components/Brokers/Broker/Configs/Configs.styled.ts index 03030b12498..cced53373e8 100644 --- a/kafka-ui-react-app/src/components/Brokers/Broker/Configs/Configs.styled.ts +++ b/kafka-ui-react-app/src/components/Brokers/Broker/Configs/Configs.styled.ts @@ -24,3 +24,13 @@ export const SearchWrapper = styled.div` margin: 10px; width: 21%; `; + +export const Source = styled.div` + display: flex; + align-content: center; + svg { + margin-left: 10px; + vertical-align: middle; + cursor: pointer; + } +`; diff --git a/kafka-ui-react-app/src/components/Brokers/Broker/Configs/Configs.tsx b/kafka-ui-react-app/src/components/Brokers/Broker/Configs/Configs.tsx index 31a42f60df4..8f909facac0 100644 --- a/kafka-ui-react-app/src/components/Brokers/Broker/Configs/Configs.tsx +++ b/kafka-ui-react-app/src/components/Brokers/Broker/Configs/Configs.tsx @@ -9,10 +9,20 @@ import { import Table from 'components/common/NewTable'; import { BrokerConfig, ConfigSource } from 'generated-sources'; import Search from 'components/common/Search/Search'; +import Tooltip from 'components/common/Tooltip/Tooltip'; +import InfoIcon from 'components/common/Icons/InfoIcon'; import InputCell from './InputCell'; import * as S from './Configs.styled'; +const tooltipContent = `DYNAMIC_TOPIC_CONFIG = dynamic topic config that is configured for a specific topic +DYNAMIC_BROKER_LOGGER_CONFIG = dynamic broker logger config that is configured for a specific broker +DYNAMIC_BROKER_CONFIG = dynamic broker config that is configured for a specific broker +DYNAMIC_DEFAULT_BROKER_CONFIG = dynamic broker config that is configured as default for all brokers in the cluster +STATIC_BROKER_CONFIG = static broker config provided as broker properties at start up (e.g. server.properties file) +DEFAULT_CONFIG = built-in default configuration for configs that have a default value +UNKNOWN = source unknown e.g. in the ConfigEntry used for alter requests where source is not set`; + const Configs: React.FC = () => { const [keyword, setKeyword] = React.useState(''); const { clusterName, brokerId } = useAppParams<ClusterBrokerParam>(); @@ -57,7 +67,19 @@ const Configs: React.FC = () => { cell: renderCell, }, { - header: 'Source', + // eslint-disable-next-line react/no-unstable-nested-components + header: () => { + return ( + <S.Source> + Source + <Tooltip + value={<InfoIcon />} + content={tooltipContent} + placement="top-end" + /> + </S.Source> + ); + }, accessorKey: 'source', }, ], diff --git a/kafka-ui-react-app/src/components/common/Icons/InfoIcon.tsx b/kafka-ui-react-app/src/components/common/Icons/InfoIcon.tsx new file mode 100644 index 00000000000..8e947c6ca97 --- /dev/null +++ b/kafka-ui-react-app/src/components/common/Icons/InfoIcon.tsx @@ -0,0 +1,53 @@ +import React from 'react'; + +const InfoIcon: React.FC = () => { + return ( + <svg + xmlns="http://www.w3.org/2000/svg" + viewBox="0 0 64 64" + width="12" + height="12" + aria-labelledby="title" + aria-describedby="desc" + role="img" + > + <desc>A line styled icon from Orion Icon Library.</desc> + <circle + data-name="layer2" + cx="32" + cy="32" + r="30" + fill="none" + stroke="#202020" + strokeMiterlimit="10" + strokeWidth="2" + strokeLinejoin="round" + strokeLinecap="round" + /> + <path + data-name="layer1" + fill="none" + stroke="#202020" + strokeMiterlimit="10" + strokeWidth="2" + d="M28 26h4v22m-4 .008h8" + strokeLinejoin="round" + strokeLinecap="round" + /> + <circle + data-name="layer1" + cx="31" + cy="19" + r="2" + fill="none" + stroke="#202020" + strokeMiterlimit="10" + strokeWidth="2" + strokeLinejoin="round" + strokeLinecap="round" + /> + </svg> + ); +}; + +export default InfoIcon; diff --git a/kafka-ui-react-app/src/components/common/Tooltip/Tooltip.styled.ts b/kafka-ui-react-app/src/components/common/Tooltip/Tooltip.styled.ts new file mode 100644 index 00000000000..ee352995bd4 --- /dev/null +++ b/kafka-ui-react-app/src/components/common/Tooltip/Tooltip.styled.ts @@ -0,0 +1,17 @@ +import styled from 'styled-components'; + +export const MessageTooltip = styled.div` + max-width: 100%; + max-height: 100%; + background-color: ${({ theme }) => theme.tooltip.bg}; + color: ${({ theme }) => theme.tooltip.text}; + border-radius: 6px; + padding: 5px; + z-index: 1; + white-space: pre-wrap; +`; + +export const Wrapper = styled.div` + display: flex; + align-items: center; +`; diff --git a/kafka-ui-react-app/src/components/common/Tooltip/Tooltip.tsx b/kafka-ui-react-app/src/components/common/Tooltip/Tooltip.tsx new file mode 100644 index 00000000000..cd76f4cc4f4 --- /dev/null +++ b/kafka-ui-react-app/src/components/common/Tooltip/Tooltip.tsx @@ -0,0 +1,47 @@ +import React, { useState } from 'react'; +import { + useFloating, + useHover, + useInteractions, + Placement, +} from '@floating-ui/react-dom-interactions'; + +import * as S from './Tooltip.styled'; + +export interface PropsTypes { + value: string | JSX.Element; + content: string; + placement?: Placement; +} + +const Tooltip: React.FC<PropsTypes> = ({ value, content, placement }) => { + const [open, setOpen] = useState(false); + const { x, y, reference, floating, strategy, context } = useFloating({ + open, + onOpenChange: setOpen, + placement, + }); + + useInteractions([useHover(context)]); + + return ( + <> + <S.Wrapper ref={reference}>{value}</S.Wrapper> + {open && ( + <S.MessageTooltip + ref={floating} + style={{ + position: strategy, + top: y ?? 0, + left: x ?? 0, + width: 'max-content', + }} + > + {content} + </S.MessageTooltip> + )} + </> + ); +}; + +export default Tooltip; diff --git a/kafka-ui-react-app/src/components/common/Tooltip/__tests__/Tooltip.spec.tsx b/kafka-ui-react-app/src/components/common/Tooltip/__tests__/Tooltip.spec.tsx new file mode 100644 index 00000000000..df0cc470a1c --- /dev/null +++ b/kafka-ui-react-app/src/components/common/Tooltip/__tests__/Tooltip.spec.tsx @@ -0,0 +1,29 @@ +import React from 'react'; +import { render } from 'lib/testHelpers'; +import { screen } from '@testing-library/react'; +import Tooltip from 'components/common/Tooltip/Tooltip'; +import userEvent from '@testing-library/user-event'; + +describe('Tooltip', () => { + const tooltipText = 'tooltip'; + const tooltipContent = 'tooltip_Content'; + + const setUpComponent = () => + render(<Tooltip value={tooltipText} content={tooltipContent} />); + + it('should render the tooltip element with its value text', () => { + setUpComponent(); + expect(screen.getByText(tooltipText)).toBeInTheDocument(); + }); + + it('should render the tooltip with default closed', () => { + setUpComponent(); + expect(screen.queryByText(tooltipContent)).not.toBeInTheDocument(); + }); + + it('should render the tooltip with and open during hover', async () => { + setUpComponent(); + await userEvent.hover(screen.getByText(tooltipText)); + expect(screen.getByText(tooltipContent)).toBeInTheDocument(); + }); +}); diff --git a/kafka-ui-react-app/src/theme/theme.ts b/kafka-ui-react-app/src/theme/theme.ts index 5a2b806a01f..020d5819ee5 100644 --- a/kafka-ui-react-app/src/theme/theme.ts +++ b/kafka-ui-react-app/src/theme/theme.ts @@ -527,6 +527,10 @@ const theme = { configList: { color: Colors.neutral[30], }, + tooltip: { + bg: Colors.neutral[70], + text: Colors.neutral[0], + }, topicsList: { color: { normal: Colors.neutral[90],
null
train
val
2022-11-01T14:10:16
"2022-09-30T09:33:01Z"
Haarolean
train
provectus/kafka-ui/2726_2776
provectus/kafka-ui
provectus/kafka-ui/2726
provectus/kafka-ui/2776
[ "keyword_pr_to_issue", "connected" ]
e87178136c86be37dbf1ff08eb74142dd53ccd99
e67d940981612753b406aa6273514c0b37c04788
[]
[]
"2022-10-18T14:25:16Z"
[ "type/bug", "good first issue", "scope/frontend", "status/accepted", "status/confirmed", "hacktoberfest" ]
Unable to update topic config in case of retain time has been changed
1. Edit any topic 2. Change "time to retain data" via **selectors** 3. "Update topic" button is still disabled
[ "kafka-ui-react-app/src/components/Topics/shared/Form/TimeToRetainBtn.tsx" ]
[ "kafka-ui-react-app/src/components/Topics/shared/Form/TimeToRetainBtn.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/shared/Form/TimeToRetainBtn.tsx b/kafka-ui-react-app/src/components/Topics/shared/Form/TimeToRetainBtn.tsx index 2364d90516a..8746eef7929 100644 --- a/kafka-ui-react-app/src/components/Topics/shared/Form/TimeToRetainBtn.tsx +++ b/kafka-ui-react-app/src/components/Topics/shared/Form/TimeToRetainBtn.tsx @@ -18,7 +18,11 @@ const TimeToRetainBtn: React.FC<Props> = ({ inputName, text, value }) => { <S.Button isActive={parseFloat(watchedValue) === value} type="button" - onClick={() => setValue(inputName, value)} + onClick={() => + setValue(inputName, value, { + shouldDirty: true, + }) + } > {text} </S.Button>
null
train
val
2022-10-25T21:52:19
"2022-10-11T09:04:56Z"
Haarolean
train
provectus/kafka-ui/2103_2781
provectus/kafka-ui
provectus/kafka-ui/2103
provectus/kafka-ui/2781
[ "connected" ]
cb8049e769452a40007c29acdfae43e5e8abe53d
aa6c3083c2c44aa2a1764ea286d5fca9a34b4f9c
[ "Hello there raferg! πŸ‘‹\n\nThank you and congratulations πŸŽ‰ for opening your very first issue in this project! πŸ’–\n\nIn case you want to claim this issue, please comment down below! We will try to get back to you as soon as we can. πŸ‘€", "Hey, thanks for raising the issue. We'll take a look." ]
[]
"2022-10-19T14:28:31Z"
[ "type/bug", "good first issue", "scope/backend", "status/accepted", "hacktoberfest" ]
Topics fail to list, or refresh with an empty list when one of the connected brokers is killed
**Describe the bug** I've been testing failure handling in my local configuration, and have discovered that when i take down one of the brokers via a docker kill command, the ui cannot list any topics. It is fine listing brokers, and consumers, but it cannot list the topics. During my test i took down brdg-broker-one and it fails to list any topics (has a spinning icon indefinitely). In a separate test i took out brdg-broker-two and it managed to list the topics, but then refreshes the page with an empty list. **Set up** docker_compose.yml: ``` networks: brdgNet: name: brdgNet services: brdg-zookeeper: image: confluentinc/cp-zookeeper:5.5.0 container_name: brdg-zookeeper logging: driver: "json-file" options: max-size: "2048m" networks: - brdgNet environment: ZOOKEEPER_CLIENT_PORT: 2181 brdg-broker-one: image: confluentinc/cp-kafka:5.5.0 container_name: brdg-broker-one logging: driver: "json-file" options: max-size: "2048m" networks: - brdgNet restart: always ports: - "19092:19092" depends_on: - brdg-zookeeper environment: KAFKA_BROKER_ID: 1 KAFKA_ZOOKEEPER_CONNECT: brdg-zookeeper:2181 KAFKA_ADVERTISED_LISTENERS: PLAINTEXT://brdg-broker-one:9092 KAFKA_OFFSETS_TOPIC_REPLICATION_FACTOR: 2 KAFKA_ADVERTISED_LISTENERS: PLAINTEXT://brdg-broker-one:9092,CONNECTIONS_FROM_HOST://localhost:19092 KAFKA_LISTENER_SECURITY_PROTOCOL_MAP: PLAINTEXT:PLAINTEXT,CONNECTIONS_FROM_HOST:PLAINTEXT KAFKA_LOG_MESSAGE_TIMESTAMP_TYPE: CreateTime KAFKA_LOG_RETENTION_HOURS: 2232 brdg-broker-two: image: confluentinc/cp-kafka:5.5.0 container_name: brdg-broker-two logging: driver: "json-file" options: max-size: "2048m" networks: - brdgNet restart: always ports: - "19093:19093" depends_on: - brdg-zookeeper environment: KAFKA_BROKER_ID: 2 KAFKA_ZOOKEEPER_CONNECT: brdg-zookeeper:2181 KAFKA_ADVERTISED_LISTENERS: PLAINTEXT://brdg-broker-two:9093 KAFKA_OFFSETS_TOPIC_REPLICATION_FACTOR: 2 KAFKA_ADVERTISED_LISTENERS: PLAINTEXT://brdg-broker-two:9093,CONNECTIONS_FROM_HOST://localhost:19093 KAFKA_LISTENER_SECURITY_PROTOCOL_MAP: PLAINTEXT:PLAINTEXT,CONNECTIONS_FROM_HOST:PLAINTEXT KAFKA_LOG_MESSAGE_TIMESTAMP_TYPE: CreateTime KAFKA_LOG_RETENTION_HOURS: 2232 brdg-broker-three: image: confluentinc/cp-kafka:5.5.0 container_name: brdg-broker-three logging: driver: "json-file" options: max-size: "2048m" networks: - brdgNet restart: always ports: - "19094:19094" depends_on: - brdg-zookeeper environment: KAFKA_BROKER_ID: 3 KAFKA_ZOOKEEPER_CONNECT: brdg-zookeeper:2181 KAFKA_ADVERTISED_LISTENERS: PLAINTEXT://brdg-broker-three:9094 KAFKA_OFFSETS_TOPIC_REPLICATION_FACTOR: 2 KAFKA_ADVERTISED_LISTENERS: PLAINTEXT://brdg-broker-three:9094,CONNECTIONS_FROM_HOST://localhost:19094 KAFKA_LISTENER_SECURITY_PROTOCOL_MAP: PLAINTEXT:PLAINTEXT,CONNECTIONS_FROM_HOST:PLAINTEXT KAFKA_LOG_MESSAGE_TIMESTAMP_TYPE: CreateTime KAFKA_LOG_RETENTION_HOURS: 2232 kafka-ui: image: provectuslabs/kafka-ui:master container_name: brdg-webui logging: driver: "json-file" options: max-size: "2048m" ports: - "8080:8080" networks: - brdgNet restart: always depends_on: - brdg-broker-one - brdg-zookeeper environment: - KAFKA_CLUSTERS_0_NAME=local - KAFKA_CLUSTERS_0_BOOTSTRAPSERVERS=brdg-broker-one:9092,brdg-broker-two:9093,brdg-broker-three:9094 - KAFKA_CLUSTERS_0_ZOOKEEPER=brdg-zookeeper:2181 ``` **Steps to Reproduce** Steps to reproduce the behavior: 1. `docker-compose -f docker-compose.yml up -d` 2. open web browser to localhost:8080, open Topics tab, Brokers tab, and Consumer tab; observe contents 3. `docker kill brdg-broker-one` or `docker kill brdg-broker-two` 4. open web browser to localhost:8080, open Topics tab, Brokers tab, and Consumer tab; observe contents (note that Topics tab doesn't load, or loads incorrectly)
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/InternalPartition.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ConsumerGroupService.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/MessagesService.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/OffsetsResetService.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/TopicsService.java" ]
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/InternalPartition.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ConsumerGroupService.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/MessagesService.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/OffsetsResetService.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/TopicsService.java" ]
[]
diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/InternalPartition.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/InternalPartition.java index 76f916cb105..f5d16e0b65d 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/InternalPartition.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/InternalPartition.java @@ -13,12 +13,12 @@ public class InternalPartition { private final int inSyncReplicasCount; private final int replicasCount; - private final long offsetMin; - private final long offsetMax; + private final Long offsetMin; + private final Long offsetMax; // from log dir - private final long segmentSize; - private final long segmentCount; + private final Long segmentSize; + private final Integer segmentCount; } diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ConsumerGroupService.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ConsumerGroupService.java index ee9a9d8a6a0..2de70c88ca2 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ConsumerGroupService.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ConsumerGroupService.java @@ -49,8 +49,8 @@ private Mono<List<InternalConsumerGroup>> getConsumerGroups( var tpsFromGroupOffsets = groupOffsetsMap.values().stream() .flatMap(v -> v.keySet().stream()) .collect(Collectors.toSet()); - // 2. getting end offsets for partitions with in committed offsets - return ac.listOffsets(tpsFromGroupOffsets, OffsetSpec.latest()) + // 2. getting end offsets for partitions with committed offsets + return ac.listOffsets(tpsFromGroupOffsets, OffsetSpec.latest(), false) .map(endOffsets -> descriptions.stream() .map(desc -> { @@ -68,7 +68,7 @@ public Mono<List<InternalTopicConsumerGroup>> getConsumerGroupsForTopic(KafkaClu String topic) { return adminClientService.get(cluster) // 1. getting topic's end offsets - .flatMap(ac -> ac.listOffsets(topic, OffsetSpec.latest()) + .flatMap(ac -> ac.listTopicOffsets(topic, OffsetSpec.latest(), false) .flatMap(endOffsets -> { var tps = new ArrayList<>(endOffsets.keySet()); // 2. getting all consumer groups diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/MessagesService.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/MessagesService.java index 1f217b1e40d..ecfeda01227 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/MessagesService.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/MessagesService.java @@ -65,8 +65,8 @@ public Mono<Void> deleteTopicMessages(KafkaCluster cluster, String topicName, private Mono<Map<TopicPartition, Long>> offsetsForDeletion(KafkaCluster cluster, String topicName, List<Integer> partitionsToInclude) { return adminClientService.get(cluster).flatMap(ac -> - ac.listOffsets(topicName, OffsetSpec.earliest()) - .zipWith(ac.listOffsets(topicName, OffsetSpec.latest()), + ac.listTopicOffsets(topicName, OffsetSpec.earliest(), true) + .zipWith(ac.listTopicOffsets(topicName, OffsetSpec.latest(), true), (start, end) -> end.entrySet().stream() .filter(e -> partitionsToInclude.isEmpty() diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/OffsetsResetService.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/OffsetsResetService.java index b2675d51be8..36b812473e1 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/OffsetsResetService.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/OffsetsResetService.java @@ -47,11 +47,12 @@ private Mono<Map<TopicPartition, Long>> offsets(ReactiveAdminClient client, @Nullable Collection<Integer> partitions, OffsetSpec spec) { if (partitions == null) { - return client.listOffsets(topic, spec); + return client.listTopicOffsets(topic, spec, true); } return client.listOffsets( partitions.stream().map(idx -> new TopicPartition(topic, idx)).collect(toSet()), - spec + spec, + true ); } @@ -84,9 +85,9 @@ public Mono<Void> resetToOffsets( .collect(toMap(e -> new TopicPartition(topic, e.getKey()), Map.Entry::getValue)); return checkGroupCondition(cluster, group).flatMap( ac -> - ac.listOffsets(partitionOffsets.keySet(), OffsetSpec.earliest()) + ac.listOffsets(partitionOffsets.keySet(), OffsetSpec.earliest(), true) .flatMap(earliest -> - ac.listOffsets(partitionOffsets.keySet(), OffsetSpec.latest()) + ac.listOffsets(partitionOffsets.keySet(), OffsetSpec.latest(), true) .map(latest -> editOffsetsBounds(partitionOffsets, earliest, latest)) .flatMap(offsetsToCommit -> resetOffsets(ac, group, offsetsToCommit))) ); diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java index 463e2d9b6c6..2504473b171 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java @@ -8,12 +8,15 @@ import com.google.common.collect.Iterators; import com.provectus.kafka.ui.exception.IllegalEntityStateException; import com.provectus.kafka.ui.exception.NotFoundException; +import com.provectus.kafka.ui.exception.ValidationException; import com.provectus.kafka.ui.util.MapUtil; import com.provectus.kafka.ui.util.NumberUtil; +import com.provectus.kafka.ui.util.annotations.KafkaClientInternalsDependant; import java.io.Closeable; import java.util.ArrayList; import java.util.Arrays; import java.util.Collection; +import java.util.HashSet; import java.util.Iterator; import java.util.List; import java.util.Map; @@ -25,6 +28,7 @@ import java.util.concurrent.atomic.AtomicInteger; import java.util.function.BiFunction; import java.util.function.Function; +import java.util.function.Predicate; import java.util.stream.Collectors; import java.util.stream.Stream; import javax.annotation.Nullable; @@ -51,6 +55,7 @@ import org.apache.kafka.common.KafkaFuture; import org.apache.kafka.common.Node; import org.apache.kafka.common.TopicPartition; +import org.apache.kafka.common.TopicPartitionInfo; import org.apache.kafka.common.TopicPartitionReplica; import org.apache.kafka.common.acl.AclOperation; import org.apache.kafka.common.config.ConfigResource; @@ -400,32 +405,81 @@ public Mono<Void> alterConsumerGroupOffsets(String groupId, Map<TopicPartition, .all()); } - public Mono<Map<TopicPartition, Long>> listOffsets(String topic, - OffsetSpec offsetSpec) { - return topicPartitions(topic).flatMap(tps -> listOffsets(tps, offsetSpec)); + /** + * List offset for the topic's partitions and OffsetSpec. + * @param failOnUnknownLeader true - throw exception in case of no-leader partitions, + * false - skip partitions with no leader + */ + public Mono<Map<TopicPartition, Long>> listTopicOffsets(String topic, + OffsetSpec offsetSpec, + boolean failOnUnknownLeader) { + return describeTopic(topic) + .map(td -> filterPartitionsWithLeaderCheck(List.of(td), p -> true, failOnUnknownLeader)) + .flatMap(partitions -> listOffsetsUnsafe(partitions, offsetSpec)); } + /** + * List offset for the specified partitions and OffsetSpec. + * @param failOnUnknownLeader true - throw exception in case of no-leader partitions, + * false - skip partitions with no leader + */ public Mono<Map<TopicPartition, Long>> listOffsets(Collection<TopicPartition> partitions, - OffsetSpec offsetSpec) { - //TODO: need to split this into multiple calls if number of target partitions is big - return toMono( - client.listOffsets(partitions.stream().collect(toMap(tp -> tp, tp -> offsetSpec))).all()) - .map(offsets -> offsets.entrySet() - .stream() - // filtering partitions for which offsets were not found - .filter(e -> e.getValue().offset() >= 0) - .collect(toMap(Map.Entry::getKey, e -> e.getValue().offset()))); - } - - private Mono<Set<TopicPartition>> topicPartitions(String topic) { - return toMono(client.describeTopics(List.of(topic)).all()) - .map(r -> r.values().stream() - .findFirst() - .stream() - .flatMap(d -> d.partitions().stream()) - .map(p -> new TopicPartition(topic, p.partition())) - .collect(Collectors.toSet()) - ); + OffsetSpec offsetSpec, + boolean failOnUnknownLeader) { + return filterPartitionsWithLeaderCheck(partitions, failOnUnknownLeader) + .flatMap(parts -> listOffsetsUnsafe(parts, offsetSpec)); + } + + private Mono<Collection<TopicPartition>> filterPartitionsWithLeaderCheck(Collection<TopicPartition> partitions, + boolean failOnUnknownLeader) { + var targetTopics = partitions.stream().map(TopicPartition::topic).collect(Collectors.toSet()); + return describeTopicsImpl(targetTopics) + .map(descriptions -> + filterPartitionsWithLeaderCheck( + descriptions.values(), partitions::contains, failOnUnknownLeader)); + } + + private Set<TopicPartition> filterPartitionsWithLeaderCheck(Collection<TopicDescription> topicDescriptions, + Predicate<TopicPartition> partitionPredicate, + boolean failOnUnknownLeader) { + var goodPartitions = new HashSet<TopicPartition>(); + for (TopicDescription description : topicDescriptions) { + for (TopicPartitionInfo partitionInfo : description.partitions()) { + TopicPartition topicPartition = new TopicPartition(description.name(), partitionInfo.partition()); + if (!partitionPredicate.test(topicPartition)) { + continue; + } + if (partitionInfo.leader() != null) { + goodPartitions.add(topicPartition); + } else if (failOnUnknownLeader) { + throw new ValidationException(String.format("Topic partition %s has no leader", topicPartition)); + } + } + } + return goodPartitions; + } + + // 1. NOTE(!): should only apply for partitions with existing leader, + // otherwise AdminClient will try to fetch topic metadata, fail and retry infinitely (until timeout) + // 2. TODO: check if it is a bug that AdminClient never throws LeaderNotAvailableException and just retrying instead + @KafkaClientInternalsDependant + public Mono<Map<TopicPartition, Long>> listOffsetsUnsafe(Collection<TopicPartition> partitions, + OffsetSpec offsetSpec) { + + Function<Collection<TopicPartition>, Mono<Map<TopicPartition, Long>>> call = + parts -> toMono( + client.listOffsets(parts.stream().collect(toMap(tp -> tp, tp -> offsetSpec))).all()) + .map(offsets -> offsets.entrySet().stream() + // filtering partitions for which offsets were not found + .filter(e -> e.getValue().offset() >= 0) + .collect(toMap(Map.Entry::getKey, e -> e.getValue().offset()))); + + return partitionCalls( + partitions, + 200, + call, + (m1, m2) -> ImmutableMap.<TopicPartition, Long>builder().putAll(m1).putAll(m2).build() + ); } public Mono<Void> updateBrokerConfigByName(Integer brokerId, String name, String value) { diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/TopicsService.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/TopicsService.java index 8badcebc360..7b08d69fd6c 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/TopicsService.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/TopicsService.java @@ -138,11 +138,15 @@ private Mono<InternalPartitionsOffsets> getPartitionOffsets(Map<String, TopicDes ReactiveAdminClient ac) { var topicPartitions = descriptions.values().stream() .flatMap(desc -> - desc.partitions().stream().map(p -> new TopicPartition(desc.name(), p.partition()))) + desc.partitions().stream() + // list offsets should only be applied to partitions with existing leader + // (see ReactiveAdminClient.listOffsetsUnsafe(..) docs) + .filter(tp -> tp.leader() != null) + .map(p -> new TopicPartition(desc.name(), p.partition()))) .collect(toList()); - return ac.listOffsets(topicPartitions, OffsetSpec.earliest()) - .zipWith(ac.listOffsets(topicPartitions, OffsetSpec.latest()), + return ac.listOffsetsUnsafe(topicPartitions, OffsetSpec.earliest()) + .zipWith(ac.listOffsetsUnsafe(topicPartitions, OffsetSpec.latest()), (earliest, latest) -> topicPartitions.stream() .filter(tp -> earliest.containsKey(tp) && latest.containsKey(tp))
null
train
val
2022-10-31T10:17:23
"2022-06-03T03:59:13Z"
raferg
train
provectus/kafka-ui/2758_2782
provectus/kafka-ui
provectus/kafka-ui/2758
provectus/kafka-ui/2782
[ "connected", "timestamp(timedelta=1.0, similarity=0.8795893534509392)" ]
a2e87cc8d5ba21e2b01f2f240926298f382db63c
f7f2e1dd1f6a146d7371e77bd497f82f21b03444
[ "Hello there vliska! πŸ‘‹\n\nThank you and congratulations πŸŽ‰ for opening your very first issue in this project! πŸ’–\n\nIn case you want to claim this issue, please comment down below! We will try to get back to you as soon as we can. πŸ‘€", "Hey, thanks for reaching out. We'll take a look" ]
[]
"2022-10-19T15:05:22Z"
[ "type/bug", "good first issue", "scope/frontend", "status/accepted", "hacktoberfest", "type/regression" ]
Wrong actuator path with custom SERVER_SERVLET_CONTEXT_PATH
**Describe the bug** When settings SERVER_SERVLET_CONTEXT_PATH to /kafka-ui, actuator is deployed with /kafka-ui/actuator/info, but application asks for /actuator/info. This is a regression from v0.4.0 (which we used before) - it works fine on v0.4.0 (maybe because v0.4.0 does not seem to be asking for /actuator/info endpoint). **Set up** Install latest version (c22b6b468c61da1d1a647352405f63cd0670e493335743de6c7f24a519264b35) with SERVER_SERVLET_CONTEXT_PATH: /kafka-ui **Steps to Reproduce** 1. go to the landing page **Expected behavior** Landing page works and asks for acutator in correct path (/kafka-ui/actuator/info). **Screenshots** Not working (c22b6b468c61da1d1a647352405f63cd0670e493335743de6c7f24a519264b35) : ![image](https://user-images.githubusercontent.com/23450695/196111995-150274fa-9415-483e-a28d-78447b8943e7.png)
[ "kafka-ui-react-app/src/lib/hooks/api/actuatorInfo.ts" ]
[ "kafka-ui-react-app/src/lib/hooks/api/actuatorInfo.ts" ]
[]
diff --git a/kafka-ui-react-app/src/lib/hooks/api/actuatorInfo.ts b/kafka-ui-react-app/src/lib/hooks/api/actuatorInfo.ts index 8373bf53d38..7e1835d9079 100644 --- a/kafka-ui-react-app/src/lib/hooks/api/actuatorInfo.ts +++ b/kafka-ui-react-app/src/lib/hooks/api/actuatorInfo.ts @@ -2,9 +2,10 @@ import { useQuery } from '@tanstack/react-query'; import { BASE_PARAMS, QUERY_REFETCH_OFF_OPTIONS } from 'lib/constants'; const fetchActuatorInfo = async () => { - const data = await fetch('/actuator/info', BASE_PARAMS).then((res) => - res.json() - ); + const data = await fetch( + `${BASE_PARAMS.basePath}/actuator/info`, + BASE_PARAMS + ).then((res) => res.json()); return data; };
null
train
val
2022-10-24T10:13:06
"2022-10-17T07:15:40Z"
vliska
train
provectus/kafka-ui/2774_2785
provectus/kafka-ui
provectus/kafka-ui/2774
provectus/kafka-ui/2785
[ "connected" ]
b0e0da4c0d1bd66bb2ca053f421dc4c7e78d78dc
40c3bcf82b7c21a367fb4112e7f8e51c843481ea
[ "Hello there Redbutton18! πŸ‘‹\n\nThank you and congratulations πŸŽ‰ for opening your very first issue in this project! πŸ’–\n\nIn case you want to claim this issue, please comment down below! We will try to get back to you as soon as we can. πŸ‘€", "Bug found \r\nhttps://github.com/provectus/kafka-ui/issues/2778" ]
[ "delete your token pls", "we have already defined element 'dotMenuBtn'\n- use it if it's the same element\n- define new element at the top of class as SelenideElement if it's another", "let's use single style for defining objects\ncreate string at the top of the class pls", "please don't use extra spaces", "seems that we don't need to set up for cases we use TOPIC_FOR_MESSAGES:\n.setTimeToRetainData(\"604800001\")\n.setMaxSizeOnDisk(\"10 GB\")\n.setMaxMessageBytes(\"1000012\")", "I'd like o use @Issue annotation to store the link to issue because it will be displayed in Allure report", "we need o waitUntilScreenReady first", "what do we wait here? screen wasn't changed", "in addition for TODO", "Done", "Done", "Refactored", "Done", "This fields removed from TOPIC_FOR_MESSAGES", "@issue annotation added", "Done", "Removed" ]
"2022-10-20T13:02:48Z"
[ "scope/AQA" ]
[e2e] TopicTests.clearMessagesForPartition : Clear Messages for partition
Autotest implentation for: https://app.qase.io/project/KAFKAUI?view=1&suite=5&case=19 Description: Checking Messages clearing for the selected Partition within Topic/Overview Preconditions: Login to Kafka-ui application Open the 'Local' section Select the 'Topics' Open the Topic profile which has messages for different partitions Stay in 'Overview' tab Steps: 1 Click on Menu icon for partition 2 Press 'Clear messages'Expected resultSuccess message about clearing should displayAll the messages for selected partition should be deleted
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Topic.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Topic.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Topic.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Topic.java index 725db0dd8dd..18763e37191 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Topic.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Topic.java @@ -6,5 +6,5 @@ @Data @Accessors(chain = true) public class Topic { - private String name, compactPolicyValue, timeToRetainData, maxSizeOnDisk, maxMessageBytes, messageKey, messageContent ; + private String name, cleanupPolicyValue, timeToRetainData, maxSizeOnDisk, maxMessageBytes, messageKey, messageContent ; } diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java index b455d310fbc..5b4f1d3c962 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java @@ -17,6 +17,8 @@ public class TopicDetails { protected SelenideElement loadingSpinner = $x("//*[contains(text(),'Loading')]"); protected SelenideElement dotMenuBtn = $$x("//button[@aria-label='Dropdown Toggle']").first(); + protected SelenideElement dotPartitionIdMenuBtn = $(By.cssSelector("button.sc-hOqruk.eYtACj")); + protected SelenideElement clearMessagesBtn = $x(("//div[contains(text(), 'Clear messages')]")); protected SelenideElement overviewTab = $x("//a[contains(text(),'Overview')]"); protected SelenideElement messagesTab = $x("//a[contains(text(),'Messages')]"); protected SelenideElement editSettingsTab = $x("//li[@role][contains(text(),'Edit settings')]"); @@ -45,6 +47,18 @@ public TopicDetails openTopicMenu(TopicMenu menu) { return this; } + @Step + public TopicDetails openDotPartitionIdMenu() { + dotPartitionIdMenuBtn.shouldBe(Condition.visible.because("dot menu invisible")).click(); + return this; + } + + @Step + public TopicDetails clickClearMessagesBtn() { + clearMessagesBtn.shouldBe(Condition.visible.because("Clear Messages invisible")).click(); + return this; + } + @Step public TopicDetails deleteTopic() { clickByJavaScript(dotMenuBtn); @@ -70,6 +84,11 @@ public boolean isContentMessageVisible(String contentMessage) { return contentMessage.matches(contentMessageTab.getText().trim()); } + @Step + public String MessageCountAmount() { + return $(By.xpath("//table[@class=\"sc-hiSbEG cvnuic\"]/tbody/tr/td[5]")).getText(); + } + private enum DotMenuHeaderItems { EDIT_SETTINGS("Edit settings"), CLEAR_MESSAGES("Clear messages"), @@ -91,6 +110,26 @@ public String toString() { } } + public enum DotPartitionIdMenu { + CLEAR_MESSAGES("Clear messages"); + + + private final String value; + + DotPartitionIdMenu(String value) { + this.value = value; + } + + public String getValue() { + return value; + } + + @Override + public String toString() { + return "DotPartitionIdMenuItems{" + "value='" + value + '\'' + '}'; + } + } + public enum TopicMenu { OVERVIEW("Overview"), MESSAGES("Messages"),
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java index bb1cfc34273..2b596a9e545 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java @@ -6,6 +6,7 @@ import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.AutomationStatus; import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.Suite; import com.provectus.kafka.ui.utilities.qaseIoUtils.enums.Status; +import io.qameta.allure.Issue; import io.qase.api.annotation.CaseId; import org.assertj.core.api.SoftAssertions; import org.junit.jupiter.api.*; @@ -14,6 +15,7 @@ import java.util.List; import static com.provectus.kafka.ui.pages.NaviSideBar.SideMenuOption.TOPICS; +import static com.provectus.kafka.ui.pages.topic.TopicDetails.DotPartitionIdMenu.CLEAR_MESSAGES; import static com.provectus.kafka.ui.settings.Source.CLUSTER_NAME; import static com.provectus.kafka.ui.utilities.FileUtils.fileToString; @@ -23,18 +25,23 @@ public class TopicTests extends BaseTest { private static final String SUITE_TITLE = "Topics"; private static final Topic TOPIC_FOR_UPDATE = new Topic() .setName("topic-to-update") - .setCompactPolicyValue("Compact") + .setCleanupPolicyValue("Compact") .setTimeToRetainData("604800001") .setMaxSizeOnDisk("20 GB") .setMaxMessageBytes("1000020") .setMessageKey(fileToString(System.getProperty("user.dir") + "/src/test/resources/producedkey.txt")) .setMessageContent(fileToString(System.getProperty("user.dir") + "/src/test/resources/testData.txt")); + private static final Topic TOPIC_FOR_MESSAGES = new Topic() + .setName("topic-with-clean-message-attribute") + .setMessageKey(fileToString(System.getProperty("user.dir") + "/src/test/resources/producedkey.txt")) + .setMessageContent(fileToString(System.getProperty("user.dir") + "/src/test/resources/testData.txt")); + private static final Topic TOPIC_FOR_DELETE = new Topic().setName("topic-to-delete"); private static final List<Topic> TOPIC_LIST = new ArrayList<>(); @BeforeAll public void beforeAll() { - TOPIC_LIST.addAll(List.of(TOPIC_FOR_UPDATE, TOPIC_FOR_DELETE)); + TOPIC_LIST.addAll(List.of(TOPIC_FOR_UPDATE, TOPIC_FOR_DELETE, TOPIC_FOR_MESSAGES)); TOPIC_LIST.forEach(topic -> apiHelper.createTopic(CLUSTER_NAME, topic.getName())); } @@ -81,7 +88,7 @@ public void updateTopic() { .openEditSettings(); topicCreateEditForm .waitUntilScreenReady() - .selectCleanupPolicy(TOPIC_FOR_UPDATE.getCompactPolicyValue()) + .selectCleanupPolicy(TOPIC_FOR_UPDATE.getCleanupPolicyValue()) .setMinInsyncReplicas(10) .setTimeToRetainDataInMs(TOPIC_FOR_UPDATE.getTimeToRetainData()) .setMaxSizeOnDiskInGB(TOPIC_FOR_UPDATE.getMaxSizeOnDisk()) @@ -98,7 +105,7 @@ public void updateTopic() { .waitUntilScreenReady() .openEditSettings(); SoftAssertions softly = new SoftAssertions(); - softly.assertThat(topicCreateEditForm.getCleanupPolicy()).as("Cleanup Policy").isEqualTo(TOPIC_FOR_UPDATE.getCompactPolicyValue()); + softly.assertThat(topicCreateEditForm.getCleanupPolicy()).as("Cleanup Policy").isEqualTo(TOPIC_FOR_UPDATE.getCleanupPolicyValue()); softly.assertThat(topicCreateEditForm.getTimeToRetain()).as("Time to retain").isEqualTo(TOPIC_FOR_UPDATE.getTimeToRetainData()); softly.assertThat(topicCreateEditForm.getMaxSizeOnDisk()).as("Max size on disk").isEqualTo(TOPIC_FOR_UPDATE.getMaxSizeOnDisk()); softly.assertThat(topicCreateEditForm.getMaxMessageBytes()).as("Max message bytes").isEqualTo(TOPIC_FOR_UPDATE.getMaxMessageBytes()); @@ -126,7 +133,7 @@ public void deleteTopic() { Assertions.assertFalse(topicsList.isTopicVisible(TOPIC_FOR_DELETE.getName()), "isTopicVisible"); TOPIC_LIST.remove(TOPIC_FOR_DELETE); } - + @DisplayName("produce message") @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) @AutomationStatus(status = Status.AUTOMATED) @@ -137,24 +144,55 @@ void produceMessage() { .openSideMenu(TOPICS); topicsList .waitUntilScreenReady() - .openTopic(TOPIC_FOR_UPDATE.getName()); + .openTopic(TOPIC_FOR_MESSAGES.getName()); topicDetails .waitUntilScreenReady() .openTopicMenu(TopicDetails.TopicMenu.MESSAGES) .clickProduceMessageBtn(); produceMessagePanel .waitUntilScreenReady() - .setContentFiled(TOPIC_FOR_UPDATE.getMessageContent()) - .setKeyField(TOPIC_FOR_UPDATE.getMessageKey()) + .setContentFiled(TOPIC_FOR_MESSAGES.getMessageContent()) + .setKeyField(TOPIC_FOR_MESSAGES.getMessageKey()) .submitProduceMessage(); topicDetails .waitUntilScreenReady(); SoftAssertions softly = new SoftAssertions(); - softly.assertThat(topicDetails.isKeyMessageVisible((TOPIC_FOR_UPDATE.getMessageKey()))).withFailMessage("isKeyMessageVisible()").isTrue(); - softly.assertThat(topicDetails.isContentMessageVisible((TOPIC_FOR_UPDATE.getMessageContent()).trim())).withFailMessage("isContentMessageVisible()").isTrue(); + softly.assertThat(topicDetails.isKeyMessageVisible((TOPIC_FOR_MESSAGES.getMessageKey()))).withFailMessage("isKeyMessageVisible()").isTrue(); + softly.assertThat(topicDetails.isContentMessageVisible((TOPIC_FOR_MESSAGES.getMessageContent()).trim())).withFailMessage("isContentMessageVisible()").isTrue(); softly.assertAll(); } + @Issue("Uncomment last assertion after bug https://github.com/provectus/kafka-ui/issues/2778 fix") + @DisplayName("clear message") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(19) + @Test + void clearMessage() { + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .openTopic(TOPIC_FOR_MESSAGES.getName()); + topicDetails + .waitUntilScreenReady() + .openTopicMenu(TopicDetails.TopicMenu.OVERVIEW) + .clickProduceMessageBtn(); + produceMessagePanel + .waitUntilScreenReady() + .setContentFiled(TOPIC_FOR_MESSAGES.getMessageContent()) + .setKeyField(TOPIC_FOR_MESSAGES.getMessageKey()) + .submitProduceMessage(); + topicDetails + .waitUntilScreenReady(); + String messageAmount = topicDetails.MessageCountAmount(); + Assertions.assertEquals(messageAmount,topicDetails.MessageCountAmount()); + topicDetails + .openDotPartitionIdMenu() + .clickClearMessagesBtn(); +// Assertions.assertEquals(Integer.toString(Integer.valueOf(messageAmount)-1),topicDetails.MessageCountAmount()); + } + @AfterAll public void afterAll() { TOPIC_LIST.forEach(topic -> apiHelper.deleteTopic(CLUSTER_NAME, topic.getName()));
train
val
2022-10-21T07:40:35
"2022-10-18T11:58:18Z"
Redbutton18
train
provectus/kafka-ui/2787_2788
provectus/kafka-ui
provectus/kafka-ui/2787
provectus/kafka-ui/2788
[ "connected" ]
81072541a8aab5271c63ae168f44e3ac86fcce16
9a3cbfa14b8864bee4cfa91c2b00e57be28ad59a
[]
[ "πŸ‘πŸΌ" ]
"2022-10-20T15:39:25Z"
[ "type/enhancement", "scope/backend", "status/accepted" ]
BE: Topic default config values are hard-coded
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/InternalTopicConfig.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/TopicsService.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/util/KafkaConstants.java", "kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml" ]
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/InternalTopicConfig.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/TopicsService.java", "kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml" ]
[ "kafka-ui-api/src/test/java/com/provectus/kafka/ui/KafkaConsumerTests.java" ]
diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/InternalTopicConfig.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/InternalTopicConfig.java index 294894ebc2f..d061dd49813 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/InternalTopicConfig.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/InternalTopicConfig.java @@ -1,8 +1,5 @@ package com.provectus.kafka.ui.model; -import static com.provectus.kafka.ui.util.KafkaConstants.TOPIC_DEFAULT_CONFIGS; -import static org.apache.kafka.common.config.TopicConfig.MESSAGE_FORMAT_VERSION_CONFIG; - import java.util.List; import lombok.Builder; import lombok.Data; @@ -19,6 +16,7 @@ public class InternalTopicConfig { private final boolean isSensitive; private final boolean isReadOnly; private final List<ConfigEntry.ConfigSynonym> synonyms; + private final String doc; public static InternalTopicConfig from(ConfigEntry configEntry) { InternalTopicConfig.InternalTopicConfigBuilder builder = InternalTopicConfig.builder() @@ -27,11 +25,22 @@ public static InternalTopicConfig from(ConfigEntry configEntry) { .source(configEntry.source()) .isReadOnly(configEntry.isReadOnly()) .isSensitive(configEntry.isSensitive()) - .synonyms(configEntry.synonyms()); - if (configEntry.name().equals(MESSAGE_FORMAT_VERSION_CONFIG)) { + .synonyms(configEntry.synonyms()) + .doc(configEntry.documentation()); + + if (configEntry.source() == ConfigEntry.ConfigSource.DEFAULT_CONFIG) { + // this is important case, because for some configs like "confluent.*" no synonyms returned, but + // they are set by default and "source" == DEFAULT_CONFIG builder.defaultValue(configEntry.value()); } else { - builder.defaultValue(TOPIC_DEFAULT_CONFIGS.get(configEntry.name())); + // normally by default first entity of synonyms values will be used. + configEntry.synonyms().stream() + // skipping DYNAMIC_TOPIC_CONFIG value - which is explicitly set value when + // topic was created (not default), see ConfigEntry.synonyms() doc + .filter(s -> s.source() != ConfigEntry.ConfigSource.DYNAMIC_TOPIC_CONFIG) + .map(ConfigEntry.ConfigSynonym::value) + .findFirst() + .ifPresent(builder::defaultValue); } return builder.build(); } diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java index c08b839410e..463e2d9b6c6 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java @@ -12,6 +12,7 @@ import com.provectus.kafka.ui.util.NumberUtil; import java.io.Closeable; import java.util.ArrayList; +import java.util.Arrays; import java.util.Collection; import java.util.Iterator; import java.util.List; @@ -69,8 +70,24 @@ public class ReactiveAdminClient implements Closeable { private enum SupportedFeature { - INCREMENTAL_ALTER_CONFIGS, - ALTER_CONFIGS + INCREMENTAL_ALTER_CONFIGS(2.3f), + CONFIG_DOCUMENTATION_RETRIEVAL(2.6f); + + private final float sinceVersion; + + SupportedFeature(float sinceVersion) { + this.sinceVersion = sinceVersion; + } + + static Set<SupportedFeature> forVersion(float kafkaVersion) { + return Arrays.stream(SupportedFeature.values()) + .filter(f -> kafkaVersion >= f.sinceVersion) + .collect(Collectors.toSet()); + } + + static Set<SupportedFeature> defaultFeatures() { + return Set.of(); + } } @Value @@ -88,18 +105,15 @@ public static Mono<ReactiveAdminClient> create(AdminClient adminClient) { new ReactiveAdminClient( adminClient, ver, - Set.of(getSupportedUpdateFeatureForVersion(ver)))); + getSupportedUpdateFeaturesForVersion(ver))); } - private static SupportedFeature getSupportedUpdateFeatureForVersion(String versionStr) { + private static Set<SupportedFeature> getSupportedUpdateFeaturesForVersion(String versionStr) { try { float version = NumberUtil.parserClusterVersion(versionStr); - return version <= 2.3f - ? SupportedFeature.ALTER_CONFIGS - : SupportedFeature.INCREMENTAL_ALTER_CONFIGS; + return SupportedFeature.forVersion(version); } catch (NumberFormatException e) { - log.info("Assuming non-incremental alter configs due to version parsing error"); - return SupportedFeature.ALTER_CONFIGS; + return SupportedFeature.defaultFeatures(); } } @@ -145,20 +159,21 @@ public String getVersion() { } public Mono<Map<String, List<ConfigEntry>>> getTopicsConfig() { - return listTopics(true).flatMap(this::getTopicsConfig); + return listTopics(true).flatMap(topics -> getTopicsConfig(topics, false)); } - public Mono<Map<String, List<ConfigEntry>>> getTopicsConfig(Collection<String> topicNames) { + public Mono<Map<String, List<ConfigEntry>>> getTopicsConfig(Collection<String> topicNames, boolean includeDoc) { + var includeDocFixed = features.contains(SupportedFeature.CONFIG_DOCUMENTATION_RETRIEVAL) && includeDoc; // we need to partition calls, because it can lead to AdminClient timeouts in case of large topics count return partitionCalls( topicNames, 200, - this::getTopicsConfigImpl, + part -> getTopicsConfigImpl(part, includeDocFixed), (m1, m2) -> ImmutableMap.<String, List<ConfigEntry>>builder().putAll(m1).putAll(m2).build() ); } - private Mono<Map<String, List<ConfigEntry>>> getTopicsConfigImpl(Collection<String> topicNames) { + private Mono<Map<String, List<ConfigEntry>>> getTopicsConfigImpl(Collection<String> topicNames, boolean includeDoc) { List<ConfigResource> resources = topicNames.stream() .map(topicName -> new ConfigResource(ConfigResource.Type.TOPIC, topicName)) .collect(toList()); @@ -166,7 +181,7 @@ private Mono<Map<String, List<ConfigEntry>>> getTopicsConfigImpl(Collection<Stri return toMonoWithExceptionFilter( client.describeConfigs( resources, - new DescribeConfigsOptions().includeSynonyms(true)).values(), + new DescribeConfigsOptions().includeSynonyms(true).includeDocumentation(includeDoc)).values(), UnknownTopicOrPartitionException.class ).map(config -> config.entrySet().stream() .collect(toMap( diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/TopicsService.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/TopicsService.java index 2dffe0de75e..8badcebc360 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/TopicsService.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/TopicsService.java @@ -68,7 +68,7 @@ public Mono<List<InternalTopic>> loadTopics(KafkaCluster c, List<String> topics) } return adminClientService.get(c) .flatMap(ac -> - ac.describeTopics(topics).zipWith(ac.getTopicsConfig(topics), + ac.describeTopics(topics).zipWith(ac.getTopicsConfig(topics, false), (descriptions, configs) -> { statisticsCache.update(c, descriptions, configs); return getPartitionOffsets(descriptions, ac).map(offsets -> { @@ -160,7 +160,7 @@ public Mono<InternalTopic> getTopicDetails(KafkaCluster cluster, String topicNam public Mono<List<ConfigEntry>> getTopicConfigs(KafkaCluster cluster, String topicName) { return adminClientService.get(cluster) - .flatMap(ac -> ac.getTopicsConfig(List.of(topicName))) + .flatMap(ac -> ac.getTopicsConfig(List.of(topicName), true)) .map(m -> m.values().stream().findFirst().orElseThrow(TopicNotFoundException::new)); } diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/util/KafkaConstants.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/util/KafkaConstants.java deleted file mode 100644 index aa482c57b59..00000000000 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/util/KafkaConstants.java +++ /dev/null @@ -1,65 +0,0 @@ -package com.provectus.kafka.ui.util; - -import static org.apache.kafka.common.config.TopicConfig.CLEANUP_POLICY_CONFIG; -import static org.apache.kafka.common.config.TopicConfig.CLEANUP_POLICY_DELETE; -import static org.apache.kafka.common.config.TopicConfig.COMPRESSION_TYPE_CONFIG; -import static org.apache.kafka.common.config.TopicConfig.DELETE_RETENTION_MS_CONFIG; -import static org.apache.kafka.common.config.TopicConfig.FILE_DELETE_DELAY_MS_CONFIG; -import static org.apache.kafka.common.config.TopicConfig.FLUSH_MESSAGES_INTERVAL_CONFIG; -import static org.apache.kafka.common.config.TopicConfig.FLUSH_MS_CONFIG; -import static org.apache.kafka.common.config.TopicConfig.INDEX_INTERVAL_BYTES_CONFIG; -import static org.apache.kafka.common.config.TopicConfig.MAX_COMPACTION_LAG_MS_CONFIG; -import static org.apache.kafka.common.config.TopicConfig.MAX_MESSAGE_BYTES_CONFIG; -import static org.apache.kafka.common.config.TopicConfig.MESSAGE_DOWNCONVERSION_ENABLE_CONFIG; -import static org.apache.kafka.common.config.TopicConfig.MESSAGE_TIMESTAMP_DIFFERENCE_MAX_MS_CONFIG; -import static org.apache.kafka.common.config.TopicConfig.MESSAGE_TIMESTAMP_TYPE_CONFIG; -import static org.apache.kafka.common.config.TopicConfig.MIN_CLEANABLE_DIRTY_RATIO_CONFIG; -import static org.apache.kafka.common.config.TopicConfig.MIN_COMPACTION_LAG_MS_CONFIG; -import static org.apache.kafka.common.config.TopicConfig.MIN_IN_SYNC_REPLICAS_CONFIG; -import static org.apache.kafka.common.config.TopicConfig.PREALLOCATE_CONFIG; -import static org.apache.kafka.common.config.TopicConfig.RETENTION_BYTES_CONFIG; -import static org.apache.kafka.common.config.TopicConfig.RETENTION_MS_CONFIG; -import static org.apache.kafka.common.config.TopicConfig.SEGMENT_BYTES_CONFIG; -import static org.apache.kafka.common.config.TopicConfig.SEGMENT_INDEX_BYTES_CONFIG; -import static org.apache.kafka.common.config.TopicConfig.SEGMENT_JITTER_MS_CONFIG; -import static org.apache.kafka.common.config.TopicConfig.SEGMENT_MS_CONFIG; -import static org.apache.kafka.common.config.TopicConfig.UNCLEAN_LEADER_ELECTION_ENABLE_CONFIG; - -import java.util.AbstractMap; -import java.util.Map; - -public final class KafkaConstants { - - private static final String LONG_MAX_STRING = Long.toString(Long.MAX_VALUE); - - public static final Map<String, String> TOPIC_DEFAULT_CONFIGS = Map.ofEntries( - new AbstractMap.SimpleEntry<>(CLEANUP_POLICY_CONFIG, CLEANUP_POLICY_DELETE), - new AbstractMap.SimpleEntry<>(COMPRESSION_TYPE_CONFIG, "producer"), - new AbstractMap.SimpleEntry<>(DELETE_RETENTION_MS_CONFIG, "86400000"), - new AbstractMap.SimpleEntry<>(FILE_DELETE_DELAY_MS_CONFIG, "60000"), - new AbstractMap.SimpleEntry<>(FLUSH_MESSAGES_INTERVAL_CONFIG, LONG_MAX_STRING), - new AbstractMap.SimpleEntry<>(FLUSH_MS_CONFIG, LONG_MAX_STRING), - new AbstractMap.SimpleEntry<>("follower.replication.throttled.replicas", ""), - new AbstractMap.SimpleEntry<>(INDEX_INTERVAL_BYTES_CONFIG, "4096"), - new AbstractMap.SimpleEntry<>("leader.replication.throttled.replicas", ""), - new AbstractMap.SimpleEntry<>(MAX_COMPACTION_LAG_MS_CONFIG, LONG_MAX_STRING), - new AbstractMap.SimpleEntry<>(MAX_MESSAGE_BYTES_CONFIG, "1000012"), - new AbstractMap.SimpleEntry<>(MESSAGE_TIMESTAMP_DIFFERENCE_MAX_MS_CONFIG, LONG_MAX_STRING), - new AbstractMap.SimpleEntry<>(MESSAGE_TIMESTAMP_TYPE_CONFIG, "CreateTime"), - new AbstractMap.SimpleEntry<>(MIN_CLEANABLE_DIRTY_RATIO_CONFIG, "0.5"), - new AbstractMap.SimpleEntry<>(MIN_COMPACTION_LAG_MS_CONFIG, "0"), - new AbstractMap.SimpleEntry<>(MIN_IN_SYNC_REPLICAS_CONFIG, "1"), - new AbstractMap.SimpleEntry<>(PREALLOCATE_CONFIG, "false"), - new AbstractMap.SimpleEntry<>(RETENTION_BYTES_CONFIG, "-1"), - new AbstractMap.SimpleEntry<>(RETENTION_MS_CONFIG, "604800000"), - new AbstractMap.SimpleEntry<>(SEGMENT_BYTES_CONFIG, "1073741824"), - new AbstractMap.SimpleEntry<>(SEGMENT_INDEX_BYTES_CONFIG, "10485760"), - new AbstractMap.SimpleEntry<>(SEGMENT_JITTER_MS_CONFIG, "0"), - new AbstractMap.SimpleEntry<>(SEGMENT_MS_CONFIG, "604800000"), - new AbstractMap.SimpleEntry<>(UNCLEAN_LEADER_ELECTION_ENABLE_CONFIG, "false"), - new AbstractMap.SimpleEntry<>(MESSAGE_DOWNCONVERSION_ENABLE_CONFIG, "true") - ); - - private KafkaConstants() { - } -} \ No newline at end of file diff --git a/kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml b/kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml index c4765fb2202..6a71863c7e8 100644 --- a/kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml +++ b/kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml @@ -2213,6 +2213,8 @@ components: type: array items: $ref: "#/components/schemas/ConfigSynonym" + doc: + type: string required: - name
diff --git a/kafka-ui-api/src/test/java/com/provectus/kafka/ui/KafkaConsumerTests.java b/kafka-ui-api/src/test/java/com/provectus/kafka/ui/KafkaConsumerTests.java index d248edf5e8c..ff11aa6656a 100644 --- a/kafka-ui-api/src/test/java/com/provectus/kafka/ui/KafkaConsumerTests.java +++ b/kafka-ui-api/src/test/java/com/provectus/kafka/ui/KafkaConsumerTests.java @@ -3,10 +3,10 @@ import static org.assertj.core.api.Assertions.assertThat; import static org.springframework.http.MediaType.TEXT_EVENT_STREAM; -import com.provectus.kafka.ui.api.model.TopicConfig; import com.provectus.kafka.ui.model.BrokerConfigDTO; import com.provectus.kafka.ui.model.PartitionsIncreaseDTO; import com.provectus.kafka.ui.model.PartitionsIncreaseResponseDTO; +import com.provectus.kafka.ui.model.TopicConfigDTO; import com.provectus.kafka.ui.model.TopicCreationDTO; import com.provectus.kafka.ui.model.TopicDetailsDTO; import com.provectus.kafka.ui.model.TopicMessageEventDTO; @@ -206,12 +206,12 @@ public void shouldRetrieveTopicConfig() { .expectStatus() .isOk(); - List<TopicConfig> configs = webTestClient.get() + List<TopicConfigDTO> configs = webTestClient.get() .uri("/api/clusters/{clusterName}/topics/{topicName}/config", LOCAL, topicName) .exchange() .expectStatus() .isOk() - .expectBodyList(TopicConfig.class) + .expectBodyList(TopicConfigDTO.class) .returnResult() .getResponseBody();
test
val
2022-10-29T20:45:21
"2022-10-20T15:33:00Z"
iliax
train
provectus/kafka-ui/2721_2804
provectus/kafka-ui
provectus/kafka-ui/2721
provectus/kafka-ui/2804
[ "keyword_pr_to_issue" ]
5a67adbf3e8fd5f71957b847a162d29fc3e5c12f
a12380100e061b55c6e4636e2899da0da38006b7
[ "I'd like to give it a try! Could I get assigned to this task please?", "Sure, please.Β On 22 Oct 2022, at 08:59, Artem Tanyhin ***@***.***> wrote:ο»Ώ\nI'd like to give it a try! Could I get assigned to this task please?\n\nβ€”Reply to this email directly, view it on GitHub, or unsubscribe.You are receiving this because you are subscribed to this thread.Message ID: ***@***.***>" ]
[]
"2022-10-23T00:40:29Z"
[ "type/bug", "good first issue", "scope/frontend", "status/accepted", "status/confirmed", "hacktoberfest" ]
KSQL: "Execute KSQL Request" Stream properties overlayed by KSQL query window
**Describe the bug** Layout of the KSQL query window is broken and Stream properties input left border is way under the KSQL query window. **Set up** [0939d61](https://github.com/provectus/kafka-ui/commit/0939d61) **Steps to Reproduce** 1. Open the KSQL tab. 2. Push the "Execute KSQL Request" button **Actual behavior** All the stream properties edit fields are under the query window (all the additional ones are as well) https://user-images.githubusercontent.com/112083452/195013920-933b7580-5f80-4d2c-afab-84ae089d6324.mov **Expected behavior** Normal window layout for every field. Including all the additional stream properties. **Additional context** Additional stream properties are also under the query window.
[ "kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/QueryForm.styled.ts" ]
[ "kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/QueryForm.styled.ts" ]
[]
diff --git a/kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/QueryForm.styled.ts b/kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/QueryForm.styled.ts index a8fa1cf7afa..3fc59fcecf4 100644 --- a/kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/QueryForm.styled.ts +++ b/kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/QueryForm.styled.ts @@ -34,17 +34,22 @@ export const StreamPropertiesContainer = styled.label` `; export const InputsContainer = styled.div` + overflow: hidden; + width: 100%; display: flex; justify-content: center; gap: 10px; `; export const StreamPropertiesInputWrapper = styled.div` + & { + width: 100%; + } & > input { + width: 100%; height: 40px; border: 1px solid grey; border-radius: 4px; - min-width: 300px; font-size: 16px; padding-left: 15px; }
null
train
val
2022-10-24T21:34:20
"2022-10-11T06:39:04Z"
BulatKha
train
provectus/kafka-ui/2675_2812
provectus/kafka-ui
provectus/kafka-ui/2675
provectus/kafka-ui/2812
[ "connected" ]
dabe2878c1610fc6631d3ab81bb7aa7048617a47
e87178136c86be37dbf1ff08eb74142dd53ccd99
[ "Ready to take this issue.", "So I started this, but not able to build the UI locally. I followed the instructions here:\r\n\r\nhttps://github.com/provectus/kafka-ui/tree/master/kafka-ui-react-app#getting-started\r\n\r\nMy node_modules get installed and I launch the docker containers successfully. The command to start runs as well, and when I go to localhost:3000, I do not see the UI built after I made the changes.\r\n\r\nAm I doing something wrong?", "Oof, what a competition, folks. I'm not sure who should I assign this to -- the first person who asked or the one who decided to start anyway? Please check out our \"up for grabs\" board/project for other issues. We don't have much frontend ones right now, but there are still some left.", "> So I started this, but not able to build the UI locally. I followed the instructions here:\r\n> \r\n> https://github.com/provectus/kafka-ui/tree/master/kafka-ui-react-app#getting-started\r\n> \r\n> My node_modules get installed and I launch the docker containers successfully. The command to start runs as well, and when I go to localhost:3000, I do not see the UI built after I made the changes.\r\n> \r\n> Am I doing something wrong?\r\n\r\nWhat do you observe there? Do you mean your changes are not reflected there? AFAIK vite does the rebuild in runtime just fine.", "Anyone's willing to work on this? ", "Me🀚", "Do I also need to change MessagesV2 ?", "Yes, please :)", "Hey, I am a little confused...\r\nAre this the following files in kafki-ui-react-app where change is to be made?\r\n`src/components/Topics/Topic/Messages/Message.tsx`\r\n`src/components/Topics/Topic/Messages/MessagesTable.tsx`\r\n`src/components/Topics/Topic/MessagesV2/ActionCell.tsx`\r\n`src/components/Topics/Topic/MessagesV2/MessagesTable.tsx`\r\n\r\nAnd sorry for being late.", "@adityabhattad2021 if I'll start looking for all the files, I'd do the same work for the second time :)\r\n\r\nThese are the places on sshots, and there might be others.\r\n\r\n![Untitled](https://user-images.githubusercontent.com/1494347/196151351-480c4a44-eac9-48fc-be80-73e68bab47d1.jpg)\r\n![Untitled2](https://user-images.githubusercontent.com/1494347/196151412-b22a51cc-fc26-43bf-a418-ae5b060cade4.jpg)\r\n\r\n", "Could we please update also in message preview ? with expanding the message \r\n\r\n<img width=\"1725\" alt=\"value\" src=\"https://user-images.githubusercontent.com/104780608/196926364-fea26357-d826-47c4-9e72-92de597f12d1.png\">\r\n", "I'll do it.", "Hey I was not able to run react app locally by following the documentation for Docker way\r\nThis was the error\r\n![Screenshot 2022-10-24 180823](https://user-images.githubusercontent.com/93488388/197530668-d7904689-11aa-4891-9c27-08c6ac4dd325.png)\r\n", "> Hey I was not able to run react app locally by following the documentation for Docker way This was the error ![Screenshot 2022-10-24 180823](https://user-images.githubusercontent.com/93488388/197530668-d7904689-11aa-4891-9c27-08c6ac4dd325.png)\r\n\r\nit seems like you don't have the docker installed?", "I did installed and setup docker.", "> I did installed and setup docker.\r\n\r\nI don't see the initial command which caused this error. ", "I have found one more content label instead of value when downloading the Kafka msg. Is this intentional (not to break JSON parsers) or is it a bug?\r\n\r\n```json\r\n{\r\n\t\"Content\": \"Lorem ipsum dolor sit amet\",\r\n\t\"Offset\": 296897,\r\n\t\"Key\": \"12345\",\r\n\t\"Partition\": 0,\r\n\t\"Headers\": {},\r\n\t\"Timestamp\": \"2023-01-17T12:00:02.437Z\"\r\n}\r\n```", "@DanielSmetanka probably a bug, I'll raise an issue" ]
[]
"2022-10-24T12:51:56Z"
[ "good first issue", "scope/frontend", "status/accepted", "type/chore", "hacktoberfest" ]
Topic messages: rename "content" to "value"
Everywhere in the messages views rename "content" to "value". It's key/value in kafka, not key/content.
[ "kafka-ui-react-app/src/components/Topics/Topic/Messages/MessageContent/MessageContent.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/MessageContent/__tests__/MessageContent.spec.tsx" ]
[ "kafka-ui-react-app/src/components/Topics/Topic/Messages/MessageContent/MessageContent.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/MessageContent/__tests__/MessageContent.spec.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessageContent/MessageContent.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessageContent/MessageContent.tsx index 6e385916b5c..85f2dcd27c9 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessageContent/MessageContent.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessageContent/MessageContent.tsx @@ -82,7 +82,7 @@ const MessageContent: React.FC<MessageContentProps> = ({ type="button" onClick={handleContentTabClick} > - Content + Value </S.Tab> <S.Tab $active={activeTab === 'headers'} @@ -108,7 +108,7 @@ const MessageContent: React.FC<MessageContentProps> = ({ </S.Metadata> <S.Metadata> - <S.MetadataLabel>Content</S.MetadataLabel> + <S.MetadataLabel>Value</S.MetadataLabel> <span> <S.MetadataValue>{messageContentFormat}</S.MetadataValue> <S.MetadataMeta> diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessageContent/__tests__/MessageContent.spec.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessageContent/__tests__/MessageContent.spec.tsx index 8159a8f7279..e7fbd2b44ab 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessageContent/__tests__/MessageContent.spec.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessageContent/__tests__/MessageContent.spec.tsx @@ -61,7 +61,7 @@ describe('MessageContent screen', () => { }); describe('when switched to display the headers', () => { - it('makes Headers tab active', () => { + it('makes headers tab active', () => { userEvent.click(screen.getByText('Headers')); expect(screen.getByText('Headers')).toHaveStyleRule( 'background-color', @@ -70,9 +70,9 @@ describe('MessageContent screen', () => { }); }); - describe('when switched to display the content', () => { - it('makes content tab active', () => { - const contentTab = screen.getAllByText('Content'); + describe('when switched to display the value', () => { + it('makes value tab active', () => { + const contentTab = screen.getAllByText('Value'); userEvent.click(contentTab[0]); expect(contentTab[0]).toHaveStyleRule( 'background-color',
null
train
val
2022-10-24T21:18:18
"2022-09-30T09:19:57Z"
Haarolean
train
provectus/kafka-ui/2697_2814
provectus/kafka-ui
provectus/kafka-ui/2697
provectus/kafka-ui/2814
[ "keyword_pr_to_issue", "connected" ]
a2e87cc8d5ba21e2b01f2f240926298f382db63c
61ac1d921ab02d89016e46408f14e566ffd2d5de
[ "Hi I would like to work on this.Would you assign me please ?", "Hey, sure. \n\n> On 4 Oct 2022, at 00:02, fMurugi ***@***.***> wrote:\n> \n> ο»Ώ\n> Hi I would like to work on this.Would you assign me please ?\n> \n> β€”\n> Reply to this email directly, view it on GitHub, or unsubscribe.\n> You are receiving this because you are subscribed to this thread.\n", "i have a question, how are you caching these results.I cant figure out.\r\n", "@iliax ?", "@fMurugi \r\nI don't know where exact FE code in placed, but if you first go to some topic page FE will call `serdes` endpoints. After that if go back to Topics list page and go to same topic again - FE wont do any `serdes` calls anymore. (hard-refresh fixing this situation Ctr - shift - R) " ]
[]
"2022-10-24T13:42:03Z"
[ "type/bug", "good first issue", "scope/frontend", "status/accepted" ]
[UI] Do not cache `serdes` endpoint results on frontend
**Description** `serdes` endpoint responses should not be cached on frontend, because its responses relates to schema-registry state, which can be changed any time. Also, endpoint should work fast, so, there should be no performance impart on calling it .
[ "kafka-ui-react-app/src/lib/hooks/api/topicMessages.tsx" ]
[ "kafka-ui-react-app/src/lib/hooks/api/topicMessages.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/lib/hooks/api/topicMessages.tsx b/kafka-ui-react-app/src/lib/hooks/api/topicMessages.tsx index 941841f11ba..1357a15ea32 100644 --- a/kafka-ui-react-app/src/lib/hooks/api/topicMessages.tsx +++ b/kafka-ui-react-app/src/lib/hooks/api/topicMessages.tsx @@ -183,11 +183,11 @@ export const useTopicMessages = ({ export function useSerdes(props: GetSerdesRequest) { const { clusterName, topicName, use } = props; + return useQuery( ['clusters', clusterName, 'topics', topicName, 'serdes', use], () => messagesApiClient.getSerdes(props), { - refetchOnMount: false, refetchOnWindowFocus: false, refetchOnReconnect: false, refetchInterval: false,
null
test
val
2022-10-24T10:13:06
"2022-10-03T13:20:42Z"
iliax
train
provectus/kafka-ui/2783_2815
provectus/kafka-ui
provectus/kafka-ui/2783
provectus/kafka-ui/2815
[ "keyword_pr_to_issue" ]
e77b91316463863160d389b0d53e9186c69b8d41
7fe999660c93e516a5e95729a358a2eafec5889e
[]
[ "can we use id submitting property of TopicForm component?" ]
"2022-10-24T15:36:04Z"
[ "type/bug", "good first issue", "scope/frontend", "status/accepted" ]
Topic creation: lock page until creation request finished
**Steps to reproduce:** 1. Go to topic creation page 2. click create topic -> page is not locked, and if topic creation is slow than it is possible to click create multiple time. **How to fix:** Lock page and show progress bar while waiting for server response cc @armenuikafka https://user-images.githubusercontent.com/702205/196905138-a9dfd03b-7a07-44c6-a261-b003f8e9016c.mov
[ "kafka-ui-react-app/src/components/Topics/New/New.tsx" ]
[ "kafka-ui-react-app/src/components/Topics/New/New.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/New/New.tsx b/kafka-ui-react-app/src/components/Topics/New/New.tsx index cf992789362..1fbd59090d2 100644 --- a/kafka-ui-react-app/src/components/Topics/New/New.tsx +++ b/kafka-ui-react-app/src/components/Topics/New/New.tsx @@ -57,7 +57,7 @@ const New: React.FC = () => { partitionCount={Number(partitionCount)} replicationFactor={Number(replicationFactor)} inSyncReplicas={Number(inSyncReplicas)} - isSubmitting={false} + isSubmitting={createTopic.isLoading} onSubmit={methods.handleSubmit(onSubmit)} /> </FormProvider>
null
train
val
2022-11-25T11:35:06
"2022-10-20T09:02:53Z"
iliax
train
provectus/kafka-ui/2810_2823
provectus/kafka-ui
provectus/kafka-ui/2810
provectus/kafka-ui/2823
[ "connected" ]
99e50f8814e10577fead44957f41db5d213a6bb4
77f789fa6c4a2471417cf0136788ad39952df808
[]
[ "please add .withFailMessage", "seems this annotations is no used in this class", "seems this annotations is no used in this class", "this class is not used in current test, will we used it in in foreseeable future?", "please remove extra space", "why do we define this element like string? can we create SelenideElement from the start?", "Done", "Done", "Done", "Yes, In next autotests implementations", "This aproach is not unique and used in line 17. Definition of this element like string, requires to wraping it by String.format that affects on element searching and as result the Test fails with error. (Π₯Π— why and it`s requires additional investigation...)", "really?", "good. let's create only classes we use in current test next time, pls", "is it button or tab? use the type in naming pls", "add also here .withFailMessage, pls", "it's not unique, but it used when we format the string with dynamic values. can't see the reason in this case", "looks like it's consumer group id. I belive the best way is to not hardcode this name and create method openConsumerGroup(String id) and use additional variable with its id in test", "don't see any difference between element search because at 28 line you use $x(thisString) as argument for isVisible() method. tested with SelenideElement -> passed. if you have some fails or unstable result need to investigate additionally", "please not hardcode consumer id in methods", "if the value used more than one time in test mb we can create local var?", "Done", "Done", "It's Id. Refactored", "Done", "Done", "you created instance var, but there's a lot of different topics we use in tests inside class, that's why abstract name is not preferred. if you want to use in on class level, let's define specific names for TOPIC_NAME and TOPIC_ID -> what topic or where we use it. mb we can create Topic object which we use in tests. but seems we can also use as local vars (inside tests), this way we understand were we use this var and scefific namin is not required", "it's not only sourceActivitiesConsumerId. first of all please use type in var names -> Locator in such cases. and also it can be used for another consumer ids, not only for sourceActivities. rename pls -> consumerIdLocator", "it's not only connectSinkPostgresActivitiesHeader. first of all please use type in var names -> Locator in such cases. and also it can be used for another consumer ids, not only for connectSinkPostgresActivities. rename pls -> consumerIdHeaderLocator", "no space before annotation and extra next line" ]
"2022-10-25T12:04:21Z"
[ "scope/QA", "scope/AQA" ]
[e2e] TopicTests.redirectToConsumerFromTopicProfile : Redirect to consumer from topic profile
Autotest implentation for: https://app.qase.io/project/KAFKAUI?view=1&suite=5&previewMode=side&case=20 Description: Checking redirection to Consumer group from Topic profile Preconditions: Login to Kafka-ui application Open the 'Local' section Select the 'Topics' Open the Topic profile Turn to Consumers tab Steps: 1 Click on Consumer Expected result: Should redirect to Consumer group profile
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/consumer/ConsumersDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/consumer/ConsumersList.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/consumer/ConsumersDetails.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/consumer/ConsumersDetails.java new file mode 100644 index 00000000000..37b020d935a --- /dev/null +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/consumer/ConsumersDetails.java @@ -0,0 +1,32 @@ +package com.provectus.kafka.ui.pages.consumer; + +import com.codeborne.selenide.Condition; +import com.codeborne.selenide.SelenideElement; +import io.qameta.allure.Step; + +import static com.codeborne.selenide.Selenide.$x; +import static com.provectus.kafka.ui.utilities.WebUtils.isVisible; + +public class ConsumersDetails { + + protected SelenideElement loadingSpinner = $x("//*[contains(text(),'Loading')]"); + protected SelenideElement topicGrid = $x("//table"); + protected String consumerIdHeaderLocator = "//h1[contains(text(),'%s')]"; + protected String topicElementLocator = "//tbody//td//a[text()='%s']"; + @Step + public ConsumersDetails waitUntilScreenReady() { + loadingSpinner.shouldBe(Condition.disappear); + topicGrid.shouldBe(Condition.visible); + return this; + } + @Step + public boolean isRedirectedConsumerTitleVisible(String consumerGroupId) { + return isVisible($x(String.format(consumerIdHeaderLocator, consumerGroupId))); + } + @Step + public boolean isTopicInConsumersDetailsVisible(String topicName) { + topicGrid.shouldBe(Condition.visible); + return isVisible($x(String.format(topicElementLocator, topicName))); + } + +} diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/consumer/ConsumersList.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/consumer/ConsumersList.java new file mode 100644 index 00000000000..0797d121955 --- /dev/null +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/consumer/ConsumersList.java @@ -0,0 +1,11 @@ +package com.provectus.kafka.ui.pages.consumer; + +import com.codeborne.selenide.SelenideElement; +import com.provectus.kafka.ui.utilities.WaitUtils; +import lombok.experimental.ExtensionMethod; + +import static com.codeborne.selenide.Selenide.$x; +public class ConsumersList { + + protected SelenideElement consumerListHeader = $x("//h1[text()='Consumers']"); +} diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java index 5b4f1d3c962..73ce320a485 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java @@ -26,6 +26,7 @@ public class TopicDetails { protected SelenideElement confirmBtn = $x("//div[@role='dialog']//button[contains(text(),'Confirm')]"); protected SelenideElement produceMessageBtn = $x("//div//button[text()='Produce Message']"); protected SelenideElement contentMessageTab = $x("//html//div[@id='root']/div/main//table//p"); + protected String consumerIdLocator = "//a[@title='%s']"; @Step public TopicDetails waitUntilScreenReady() { @@ -74,6 +75,12 @@ public TopicDetails clickProduceMessageBtn() { return this; } + @Step + public TopicDetails openConsumerGroup(String consumerId) { + $x(String.format(consumerIdLocator, consumerId)).click(); + return this; + } + @Step public boolean isKeyMessageVisible(String keyMessage) { return keyMessage.equals($("td[title]").getText());
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java index 8d3aea55d4f..ac223c62b11 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java @@ -5,6 +5,8 @@ import com.provectus.kafka.ui.pages.connector.ConnectorCreateForm; import com.provectus.kafka.ui.pages.connector.ConnectorDetails; import com.provectus.kafka.ui.pages.connector.KafkaConnectList; +import com.provectus.kafka.ui.pages.consumer.ConsumersDetails; +import com.provectus.kafka.ui.pages.consumer.ConsumersList; import com.provectus.kafka.ui.pages.schema.SchemaCreateForm; import com.provectus.kafka.ui.pages.schema.SchemaDetails; import com.provectus.kafka.ui.pages.schema.SchemaRegistryList; @@ -25,5 +27,7 @@ public abstract class Facade { protected TopicCreateEditForm topicCreateEditForm = new TopicCreateEditForm(); protected TopicsList topicsList = new TopicsList(); protected TopicDetails topicDetails = new TopicDetails(); + protected ConsumersDetails consumersDetails = new ConsumersDetails(); + protected ConsumersList consumersList = new ConsumersList(); protected NaviSideBar naviSideBar = new NaviSideBar(); } diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java index 2b596a9e545..33ad3d015a1 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java @@ -15,14 +15,15 @@ import java.util.List; import static com.provectus.kafka.ui.pages.NaviSideBar.SideMenuOption.TOPICS; -import static com.provectus.kafka.ui.pages.topic.TopicDetails.DotPartitionIdMenu.CLEAR_MESSAGES; import static com.provectus.kafka.ui.settings.Source.CLUSTER_NAME; import static com.provectus.kafka.ui.utilities.FileUtils.fileToString; +import static org.assertj.core.api.Assertions.assertThat; @TestInstance(TestInstance.Lifecycle.PER_CLASS) public class TopicTests extends BaseTest { private static final long SUITE_ID = 2; private static final String SUITE_TITLE = "Topics"; + private static final Topic TOPIC_FOR_UPDATE = new Topic() .setName("topic-to-update") .setCleanupPolicyValue("Compact") @@ -186,11 +187,38 @@ void clearMessage() { topicDetails .waitUntilScreenReady(); String messageAmount = topicDetails.MessageCountAmount(); - Assertions.assertEquals(messageAmount,topicDetails.MessageCountAmount()); + assertThat(messageAmount) + .withFailMessage("message amount not equals").isEqualTo(topicDetails.MessageCountAmount()); topicDetails .openDotPartitionIdMenu() .clickClearMessagesBtn(); -// Assertions.assertEquals(Integer.toString(Integer.valueOf(messageAmount)-1),topicDetails.MessageCountAmount()); +// assertThat(Integer.toString(Integer.valueOf(messageAmount)-1)) +// .withFailMessage("message amount not decrease by one").isEqualTo(topicDetails.MessageCountAmount()); + } + + @DisplayName("Redirect to consumer from topic profile") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(20) + @Test + void redirectToConsumerFromTopic() { + String topicName = "source-activities"; + String consumerGroupId = "connect-sink_postgres_activities"; + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .openTopic(topicName); + topicDetails + .waitUntilScreenReady() + .openTopicMenu(TopicDetails.TopicMenu.CONSUMERS) + .openConsumerGroup(consumerGroupId); + consumersDetails + .waitUntilScreenReady(); + assertThat(consumersDetails.isRedirectedConsumerTitleVisible(consumerGroupId)) + .withFailMessage("isRedirectedConsumerTitleVisible").isTrue(); + assertThat(consumersDetails.isTopicInConsumersDetailsVisible(topicName)) + .withFailMessage("isTopicInConsumersDetailsVisible").isTrue(); } @AfterAll
train
val
2022-10-31T15:05:28
"2022-10-24T09:31:52Z"
Redbutton18
train
provectus/kafka-ui/2616_2833
provectus/kafka-ui
provectus/kafka-ui/2616
provectus/kafka-ui/2833
[ "keyword_pr_to_issue" ]
c4bbbbe3da06779907a57b05354bcfcb9919ce81
e26f9787d842ad56aefbc98c7ec6f9851fd35c1b
[ "Hello there djboris9! πŸ‘‹\n\nThank you and congratulations πŸŽ‰ for opening your very first issue in this project! πŸ’–\n\nIn case you want to claim this issue, please comment down below! We will try to get back to you as soon as we can. πŸ‘€", "Hey, thanks for reaching out.\r\n\r\nWe'll add this to our backlog :)", "Working on this in #2833 " ]
[]
"2022-10-27T13:11:34Z"
[ "type/enhancement", "good first issue", "scope/backend", "status/accepted" ]
Support passing properties to Kafka Connect client
### Which version of the app are you running? Tag: `v0.4.0` and `master@d9e74deb287e7b81172c6988dd0f99b1a97d9036` ### Is your proposal related to a problem? Currently it's not possible to configure a Kafka Connect component that has TLS Client Auth required but there could be also other settings that need to be passed to the Kafka Connect Client. ### Describe the solution you'd like I would like to propose adding a `Properties` field to the `model.KafkaConnectCluster` class like the `model.KafkaCluster` and use the properties in the client instantiation. Therefore the Kafka Connect component could be configured in the following way to allow TLS Client Auth: ```plaintext KAFKA_CLUSTERS_0_KAFKACONNECT_0_PROPERTIES_SECURITY_PROTOCOL=SSL KAFKA_CLUSTERS_0_KAFKACONNECT_0_PROPERTIES_SSL_TRUSTSTORE_LOCATION=ca.jks KAFKA_CLUSTERS_0_KAFKACONNECT_0_PROPERTIES_SSL_TRUSTSTORE_PASSWORD=xxxx KAFKA_CLUSTERS_0_KAFKACONNECT_0_PROPERTIES_SSL_KEYSTORE_LOCATION=foobar.jks KAFKA_CLUSTERS_0_KAFKACONNECT_0_PROPERTIES_SSL_KEYSTORE_PASSWORD=xxxx ``` ### Describe alternatives you've considered N/A ### Additional context This issue is propably related to #930. If I get some spare time, I will have a look by myself at it and create a PR.
[ "README.md", "documentation/compose/ssl/generate_certs.sh", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/client/RetryingKafkaConnectClient.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/config/ClustersProperties.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/KafkaConnectCluster.java" ]
[ "README.md", "documentation/compose/kafka-ssl-components.yaml", "documentation/compose/ssl/generate_certs.sh", "documentation/compose/ssl/san.cnf", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/client/RetryingKafkaConnectClient.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/config/ClustersProperties.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/KafkaConnectCluster.java" ]
[]
diff --git a/README.md b/README.md index ab829328967..3bc6e4635d9 100644 --- a/README.md +++ b/README.md @@ -196,6 +196,10 @@ For example, if you want to use an environment variable to set the `name` parame |`KAFKA_CLUSTERS_0_KAFKACONNECT_0_ADDRESS` |Address of the Kafka Connect service endpoint |`KAFKA_CLUSTERS_0_KAFKACONNECT_0_USERNAME`| Kafka Connect cluster's basic authentication username |`KAFKA_CLUSTERS_0_KAFKACONNECT_0_PASSWORD`| Kafka Connect cluster's basic authentication password +|`KAFKA_CLUSTERS_0_KAFKACONNECT_0_KEYSTORELOCATION`| Path to the JKS keystore to communicate to Kafka Connect +|`KAFKA_CLUSTERS_0_KAFKACONNECT_0_KEYSTOREPASSWORD`| Password of the JKS keystore for Kafka Connect +|`KAFKA_CLUSTERS_0_KAFKACONNECT_0_TRUSTSTORELOCATION`| Path to the JKS truststore to communicate to Kafka Connect +|`KAFKA_CLUSTERS_0_KAFKACONNECT_0_TRUSTSTOREPASSWORD`| Password of the JKS truststore for Kafka Connect |`KAFKA_CLUSTERS_0_METRICS_SSL` |Enable SSL for Metrics? `true` or `false`. For advanced setup, see `kafka-ui-jmx-secured.yml` |`KAFKA_CLUSTERS_0_METRICS_USERNAME` |Username for Metrics authentication |`KAFKA_CLUSTERS_0_METRICS_PASSWORD` |Password for Metrics authentication diff --git a/documentation/compose/kafka-ssl-components.yaml b/documentation/compose/kafka-ssl-components.yaml new file mode 100644 index 00000000000..ef1f3f88821 --- /dev/null +++ b/documentation/compose/kafka-ssl-components.yaml @@ -0,0 +1,144 @@ +--- +version: '3.4' +services: + kafka-ui: + container_name: kafka-ui + image: provectuslabs/kafka-ui:latest + ports: + - 8080:8080 + depends_on: + - kafka0 + - schemaregistry0 + - kafka-connect0 + environment: + KAFKA_CLUSTERS_0_NAME: local + KAFKA_CLUSTERS_0_PROPERTIES_SECURITY_PROTOCOL: SSL + KAFKA_CLUSTERS_0_BOOTSTRAPSERVERS: kafka0:29092 # SSL LISTENER! + KAFKA_CLUSTERS_0_PROPERTIES_SSL_TRUSTSTORE_LOCATION: /kafka.truststore.jks + KAFKA_CLUSTERS_0_PROPERTIES_SSL_TRUSTSTORE_PASSWORD: secret + KAFKA_CLUSTERS_0_PROPERTIES_SSL_KEYSTORE_LOCATION: /kafka.keystore.jks + KAFKA_CLUSTERS_0_PROPERTIES_SSL_KEYSTORE_PASSWORD: secret + KAFKA_CLUSTERS_0_PROPERTIES_SSL_ENDPOINT_IDENTIFICATION_ALGORITHM: '' # DISABLE COMMON NAME VERIFICATION + KAFKA_CLUSTERS_0_SCHEMAREGISTRY: https://schemaregistry0:8085 + KAFKA_CLUSTERS_0_SCHEMAREGISTRYSSL_KEYSTORELOCATION: /kafka.keystore.jks + KAFKA_CLUSTERS_0_SCHEMAREGISTRYSSL_KEYSTOREPASSWORD: "secret" + KAFKA_CLUSTERS_0_SCHEMAREGISTRYSSL_TRUSTSTORELOCATION: /kafka.truststore.jks + KAFKA_CLUSTERS_0_SCHEMAREGISTRYSSL_TRUSTSTOREPASSWORD: "secret" + KAFKA_CLUSTERS_0_KAFKACONNECT_0_NAME: local + KAFKA_CLUSTERS_0_KAFKACONNECT_0_ADDRESS: https://kafka-connect0:8083 + KAFKA_CLUSTERS_0_KAFKACONNECT_0_KEYSTORELOCATION: /kafka.keystore.jks + KAFKA_CLUSTERS_0_KAFKACONNECT_0_KEYSTOREPASSWORD: "secret" + KAFKA_CLUSTERS_0_KAFKACONNECT_0_TRUSTSTORELOCATION: /kafka.truststore.jks + KAFKA_CLUSTERS_0_KAFKACONNECT_0_TRUSTSTOREPASSWORD: "secret" + volumes: + - ./ssl/kafka.truststore.jks:/kafka.truststore.jks + - ./ssl/kafka.keystore.jks:/kafka.keystore.jks + + kafka0: + image: confluentinc/cp-kafka:7.2.1 + hostname: kafka0 + container_name: kafka0 + ports: + - "9092:9092" + - "9997:9997" + environment: + KAFKA_BROKER_ID: 1 + KAFKA_LISTENER_SECURITY_PROTOCOL_MAP: 'CONTROLLER:PLAINTEXT,SSL:SSL,PLAINTEXT_HOST:PLAINTEXT' + KAFKA_ADVERTISED_LISTENERS: 'SSL://kafka0:29092,PLAINTEXT_HOST://localhost:9092' + KAFKA_OFFSETS_TOPIC_REPLICATION_FACTOR: 1 + KAFKA_GROUP_INITIAL_REBALANCE_DELAY_MS: 0 + KAFKA_TRANSACTION_STATE_LOG_MIN_ISR: 1 + KAFKA_TRANSACTION_STATE_LOG_REPLICATION_FACTOR: 1 + KAFKA_JMX_PORT: 9997 + KAFKA_JMX_HOSTNAME: localhost + KAFKA_PROCESS_ROLES: 'broker,controller' + KAFKA_NODE_ID: 1 + KAFKA_CONTROLLER_QUORUM_VOTERS: '1@kafka0:29093' + KAFKA_LISTENERS: 'SSL://kafka0:29092,CONTROLLER://kafka0:29093,PLAINTEXT_HOST://0.0.0.0:9092' + KAFKA_INTER_BROKER_LISTENER_NAME: 'SSL' + KAFKA_CONTROLLER_LISTENER_NAMES: 'CONTROLLER' + KAFKA_LOG_DIRS: '/tmp/kraft-combined-logs' + KAFKA_SECURITY_PROTOCOL: SSL + KAFKA_SSL_ENABLED_MECHANISMS: PLAIN,SSL + KAFKA_SSL_KEYSTORE_FILENAME: kafka.keystore.jks + KAFKA_SSL_KEYSTORE_CREDENTIALS: creds + KAFKA_SSL_KEY_CREDENTIALS: creds + KAFKA_SSL_TRUSTSTORE_FILENAME: kafka.truststore.jks + KAFKA_SSL_TRUSTSTORE_CREDENTIALS: creds + #KAFKA_SSL_CLIENT_AUTH: 'required' + KAFKA_SSL_CLIENT_AUTH: 'requested' + KAFKA_SSL_ENDPOINT_IDENTIFICATION_ALGORITHM: '' # COMMON NAME VERIFICATION IS DISABLED SERVER-SIDE + volumes: + - ./scripts/update_run.sh:/tmp/update_run.sh + - ./ssl/creds:/etc/kafka/secrets/creds + - ./ssl/kafka.truststore.jks:/etc/kafka/secrets/kafka.truststore.jks + - ./ssl/kafka.keystore.jks:/etc/kafka/secrets/kafka.keystore.jks + command: "bash -c 'if [ ! -f /tmp/update_run.sh ]; then echo \"ERROR: Did you forget the update_run.sh file that came with this docker-compose.yml file?\" && exit 1 ; else /tmp/update_run.sh && /etc/confluent/docker/run ; fi'" + + schemaregistry0: + image: confluentinc/cp-schema-registry:7.2.1 + depends_on: + - kafka0 + environment: + SCHEMA_REGISTRY_KAFKASTORE_BOOTSTRAP_SERVERS: SSL://kafka0:29092 + SCHEMA_REGISTRY_KAFKASTORE_SECURITY_PROTOCOL: SSL + SCHEMA_REGISTRY_KAFKASTORE_SSL_TRUSTSTORE_LOCATION: /kafka.truststore.jks + SCHEMA_REGISTRY_KAFKASTORE_SSL_TRUSTSTORE_PASSWORD: secret + SCHEMA_REGISTRY_KAFKASTORE_SSL_KEYSTORE_LOCATION: /kafka.keystore.jks + SCHEMA_REGISTRY_KAFKASTORE_SSL_KEYSTORE_PASSWORD: secret + SCHEMA_REGISTRY_KAFKASTORE_SSL_KEY_PASSWORD: secret + SCHEMA_REGISTRY_HOST_NAME: schemaregistry0 + SCHEMA_REGISTRY_LISTENERS: https://schemaregistry0:8085 + SCHEMA_REGISTRY_INTER_INSTANCE_PROTOCOL: https + + SCHEMA_REGISTRY_SCHEMA_REGISTRY_INTER_INSTANCE_PROTOCOL: "https" + SCHEMA_REGISTRY_LOG4J_ROOT_LOGLEVEL: INFO + SCHEMA_REGISTRY_KAFKASTORE_TOPIC: _schemas + SCHEMA_REGISTRY_SSL_CLIENT_AUTHENTICATION: "REQUIRED" + SCHEMA_REGISTRY_SSL_TRUSTSTORE_LOCATION: /kafka.truststore.jks + SCHEMA_REGISTRY_SSL_TRUSTSTORE_PASSWORD: secret + SCHEMA_REGISTRY_SSL_KEYSTORE_LOCATION: /kafka.keystore.jks + SCHEMA_REGISTRY_SSL_KEYSTORE_PASSWORD: secret + SCHEMA_REGISTRY_SSL_KEY_PASSWORD: secret + ports: + - 8085:8085 + volumes: + - ./ssl/kafka.truststore.jks:/kafka.truststore.jks + - ./ssl/kafka.keystore.jks:/kafka.keystore.jks + + kafka-connect0: + image: confluentinc/cp-kafka-connect:7.2.1 + ports: + - 8083:8083 + depends_on: + - kafka0 + - schemaregistry0 + environment: + CONNECT_BOOTSTRAP_SERVERS: kafka0:29092 + CONNECT_GROUP_ID: compose-connect-group + CONNECT_CONFIG_STORAGE_TOPIC: _connect_configs + CONNECT_CONFIG_STORAGE_REPLICATION_FACTOR: 1 + CONNECT_OFFSET_STORAGE_TOPIC: _connect_offset + CONNECT_OFFSET_STORAGE_REPLICATION_FACTOR: 1 + CONNECT_STATUS_STORAGE_TOPIC: _connect_status + CONNECT_STATUS_STORAGE_REPLICATION_FACTOR: 1 + CONNECT_KEY_CONVERTER: org.apache.kafka.connect.storage.StringConverter + CONNECT_KEY_CONVERTER_SCHEMA_REGISTRY_URL: https://schemaregistry0:8085 + CONNECT_VALUE_CONVERTER: org.apache.kafka.connect.storage.StringConverter + CONNECT_VALUE_CONVERTER_SCHEMA_REGISTRY_URL: https://schemaregistry0:8085 + CONNECT_INTERNAL_KEY_CONVERTER: org.apache.kafka.connect.json.JsonConverter + CONNECT_INTERNAL_VALUE_CONVERTER: org.apache.kafka.connect.json.JsonConverter + CONNECT_REST_ADVERTISED_HOST_NAME: kafka-connect0 + CONNECT_PLUGIN_PATH: "/usr/share/java,/usr/share/confluent-hub-components" + CONNECT_SECURITY_PROTOCOL: "SSL" + CONNECT_SSL_KEYSTORE_LOCATION: "/kafka.keystore.jks" + CONNECT_SSL_KEY_PASSWORD: "secret" + CONNECT_SSL_KEYSTORE_PASSWORD: "secret" + CONNECT_SSL_TRUSTSTORE_LOCATION: "/kafka.truststore.jks" + CONNECT_SSL_TRUSTSTORE_PASSWORD: "secret" + CONNECT_SSL_CLIENT_AUTH: "requested" + CONNECT_REST_ADVERTISED_LISTENER: "https" + CONNECT_LISTENERS: "https://kafka-connect0:8083" + volumes: + - ./ssl/kafka.truststore.jks:/kafka.truststore.jks + - ./ssl/kafka.keystore.jks:/kafka.keystore.jks diff --git a/documentation/compose/ssl/generate_certs.sh b/documentation/compose/ssl/generate_certs.sh old mode 100644 new mode 100755 index ebb916657bd..455321ef580 --- a/documentation/compose/ssl/generate_certs.sh +++ b/documentation/compose/ssl/generate_certs.sh @@ -144,7 +144,8 @@ echo "Now the trust store's private key (CA) will sign the keystore's certificat echo openssl x509 -req -CA $CA_CERT_FILE -CAkey $trust_store_private_key_file \ -in $KEYSTORE_SIGN_REQUEST -out $KEYSTORE_SIGNED_CERT \ - -days $VALIDITY_IN_DAYS -CAcreateserial + -days $VALIDITY_IN_DAYS -CAcreateserial \ + -extensions kafka -extfile san.cnf # creates $KEYSTORE_SIGN_REQUEST_SRL which is never used or needed. echo diff --git a/documentation/compose/ssl/san.cnf b/documentation/compose/ssl/san.cnf new file mode 100644 index 00000000000..53c52a9718f --- /dev/null +++ b/documentation/compose/ssl/san.cnf @@ -0,0 +1,2 @@ +[kafka] +subjectAltName = DNS:kafka0,DNS:schemaregistry0,DNS:kafka-connect0 diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/client/RetryingKafkaConnectClient.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/client/RetryingKafkaConnectClient.java index b1115d0eef1..6f71441d177 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/client/RetryingKafkaConnectClient.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/client/RetryingKafkaConnectClient.java @@ -10,21 +10,32 @@ import com.provectus.kafka.ui.connect.model.NewConnector; import com.provectus.kafka.ui.exception.KafkaConnectConflictReponseException; import com.provectus.kafka.ui.exception.ValidationException; +import com.provectus.kafka.ui.model.InternalSchemaRegistry; +import com.provectus.kafka.ui.model.KafkaCluster; import com.provectus.kafka.ui.model.KafkaConnectCluster; +import com.provectus.kafka.ui.util.SecuredWebClient; +import io.netty.handler.ssl.SslContext; +import io.netty.handler.ssl.SslContextBuilder; +import java.io.FileInputStream; +import java.security.KeyStore; import java.text.DateFormat; import java.time.Duration; import java.util.List; import java.util.Map; import java.util.TimeZone; +import javax.net.ssl.KeyManagerFactory; +import javax.net.ssl.TrustManagerFactory; import lombok.extern.slf4j.Slf4j; import org.openapitools.jackson.nullable.JsonNullableModule; import org.springframework.core.ParameterizedTypeReference; import org.springframework.http.HttpHeaders; import org.springframework.http.HttpMethod; import org.springframework.http.MediaType; +import org.springframework.http.client.reactive.ReactorClientHttpConnector; import org.springframework.http.codec.json.Jackson2JsonDecoder; import org.springframework.http.codec.json.Jackson2JsonEncoder; import org.springframework.util.MultiValueMap; +import org.springframework.util.ResourceUtils; import org.springframework.util.unit.DataSize; import org.springframework.web.client.RestClientException; import org.springframework.web.reactive.function.client.ExchangeStrategies; @@ -32,6 +43,7 @@ import org.springframework.web.reactive.function.client.WebClientResponseException; import reactor.core.publisher.Flux; import reactor.core.publisher.Mono; +import reactor.netty.http.client.HttpClient; import reactor.util.retry.Retry; @Slf4j @@ -89,7 +101,7 @@ private static class RetryingApiClient extends ApiClient { private static final ObjectMapper mapper = buildObjectMapper(dateFormat); public RetryingApiClient(KafkaConnectCluster config, DataSize maxBuffSize) { - super(buildWebClient(mapper, maxBuffSize), mapper, dateFormat); + super(buildWebClient(mapper, maxBuffSize, config), mapper, dateFormat); setBasePath(config.getAddress()); setUsername(config.getUserName()); setPassword(config.getPassword()); @@ -101,7 +113,7 @@ public static DateFormat getDefaultDateFormat() { return dateFormat; } - public static WebClient buildWebClient(ObjectMapper mapper, DataSize maxBuffSize) { + public static WebClient buildWebClient(ObjectMapper mapper, DataSize maxBuffSize, KafkaConnectCluster config) { ExchangeStrategies strategies = ExchangeStrategies .builder() .codecs(clientDefaultCodecsConfigurer -> { @@ -113,8 +125,20 @@ public static WebClient buildWebClient(ObjectMapper mapper, DataSize maxBuffSize .maxInMemorySize((int) maxBuffSize.toBytes()); }) .build(); - WebClient.Builder webClient = WebClient.builder().exchangeStrategies(strategies); - return webClient.build(); + + try { + WebClient.Builder webClient = SecuredWebClient.configure( + config.getKeystoreLocation(), + config.getKeystorePassword(), + config.getTruststoreLocation(), + config.getTruststorePassword() + ); + + return webClient.exchangeStrategies(strategies).build(); + } catch (Exception e) { + throw new IllegalStateException( + "cannot create TLS configuration for kafka-connect cluster " + config.getName(), e); + } } public static ObjectMapper buildObjectMapper(DateFormat dateFormat) { diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/config/ClustersProperties.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/config/ClustersProperties.java index 4313c3248ec..45800278b84 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/config/ClustersProperties.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/config/ClustersProperties.java @@ -55,6 +55,10 @@ public static class ConnectCluster { String address; String userName; String password; + String keystoreLocation; + String keystorePassword; + String truststoreLocation; + String truststorePassword; } @Data diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/KafkaConnectCluster.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/KafkaConnectCluster.java index 6131f3fa9e9..4a95b0e890e 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/KafkaConnectCluster.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/KafkaConnectCluster.java @@ -13,4 +13,9 @@ public class KafkaConnectCluster { private final String address; private final String userName; private final String password; + + private final String keystoreLocation; + private final String truststoreLocation; + private final String keystorePassword; + private final String truststorePassword; }
null
train
val
2022-11-03T18:16:04
"2022-09-20T11:03:21Z"
djboris9
train
provectus/kafka-ui/2685_2834
provectus/kafka-ui
provectus/kafka-ui/2685
provectus/kafka-ui/2834
[ "keyword_pr_to_issue" ]
5a06d659e28c341dede029da982afd5ffec542c4
74b20d5a5883b14003a889816418496ea0ca47e3
[]
[]
"2022-10-27T20:16:58Z"
[ "scope/frontend", "type/refactoring", "status/accepted" ]
Migrate Topic Settings page to the new version of table component
### Describe the solution you'd like Migrate `src/components/Topics/Topic/Settings/Settings.tsx` to new version of tables `src/components/common/NewTable/Table.tsx` ### Additional context Topics -> Topic -> Settings
[ "kafka-ui-react-app/src/components/Topics/Topic/Settings/ConfigListItem.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Settings/Settings.styled.ts", "kafka-ui-react-app/src/components/Topics/Topic/Settings/Settings.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Settings/__test__/ConfigListItem.spec.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Settings/__test__/Settings.spec.tsx" ]
[ "kafka-ui-react-app/src/components/Topics/Topic/Settings/Settings.styled.ts", "kafka-ui-react-app/src/components/Topics/Topic/Settings/Settings.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Settings/__test__/Settings.spec.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Settings/ConfigListItem.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Settings/ConfigListItem.tsx deleted file mode 100644 index cc0b36caa05..00000000000 --- a/kafka-ui-react-app/src/components/Topics/Topic/Settings/ConfigListItem.tsx +++ /dev/null @@ -1,24 +0,0 @@ -import { TopicConfig } from 'generated-sources'; -import React from 'react'; - -import * as S from './Settings.styled'; - -export interface ListItemProps { - config: TopicConfig; -} - -const ConfigListItem: React.FC<ListItemProps> = ({ - config: { name, value, defaultValue }, -}) => { - const hasCustomValue = !!defaultValue && value !== defaultValue; - - return ( - <S.Row $hasCustomValue={hasCustomValue}> - <td>{name}</td> - <td>{value}</td> - <td title="Default Value">{hasCustomValue && defaultValue}</td> - </S.Row> - ); -}; - -export default ConfigListItem; diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Settings/Settings.styled.ts b/kafka-ui-react-app/src/components/Topics/Topic/Settings/Settings.styled.ts index 30aa0f2afe6..e11a2ba80b4 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Settings/Settings.styled.ts +++ b/kafka-ui-react-app/src/components/Topics/Topic/Settings/Settings.styled.ts @@ -1,13 +1,14 @@ import styled, { css } from 'styled-components'; -export const Row = styled.tr<{ $hasCustomValue?: boolean }>( - ({ theme, $hasCustomValue }) => css` - & > td { - font-weight: ${$hasCustomValue ? 500 : 400}; - &:last-child { - color: ${theme.configList.color}; - font-weight: 400; - } - } +export const Value = styled.span<{ $hasCustomValue?: boolean }>( + ({ $hasCustomValue }) => css` + font-weight: ${$hasCustomValue ? 500 : 400}; + ` +); + +export const DefaultValue = styled.span( + ({ theme }) => css` + color: ${theme.configList.color}; + font-weight: 400; ` ); diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Settings/Settings.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Settings/Settings.tsx index d7b553af9d5..001fc2c0090 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Settings/Settings.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Settings/Settings.tsx @@ -1,31 +1,62 @@ import React from 'react'; -import { Table } from 'components/common/table/Table/Table.styled'; -import TableHeaderCell from 'components/common/table/TableHeaderCell/TableHeaderCell'; +import Table from 'components/common/NewTable'; import { RouteParamsClusterTopic } from 'lib/paths'; import useAppParams from 'lib/hooks/useAppParams'; import { useTopicConfig } from 'lib/hooks/api/topics'; +import { CellContext, ColumnDef } from '@tanstack/react-table'; +import { TopicConfig } from 'generated-sources'; -import ConfigListItem from './ConfigListItem'; +import * as S from './Settings.styled'; + +const ValueCell: React.FC<CellContext<TopicConfig, unknown>> = ({ + row, + renderValue, +}) => { + const { defaultValue } = row.original; + const { value } = row.original; + const hasCustomValue = !!defaultValue && value !== defaultValue; + + return ( + <S.Value $hasCustomValue={hasCustomValue}>{renderValue<string>()}</S.Value> + ); +}; + +const DefaultValueCell: React.FC<CellContext<TopicConfig, unknown>> = ({ + row, + getValue, +}) => { + const defaultValue = getValue<TopicConfig['defaultValue']>(); + const { value } = row.original; + const hasCustomValue = !!defaultValue && value !== defaultValue; + return <S.DefaultValue>{hasCustomValue && defaultValue}</S.DefaultValue>; +}; const Settings: React.FC = () => { const props = useAppParams<RouteParamsClusterTopic>(); - const { data } = useTopicConfig(props); - return ( - <Table isFullwidth> - <thead> - <tr> - <TableHeaderCell title="Key" /> - <TableHeaderCell title="Value" /> - <TableHeaderCell title="Default Value" /> - </tr> - </thead> - <tbody> - {data?.map((item) => ( - <ConfigListItem key={item.name} config={item} /> - ))} - </tbody> - </Table> + const { data = [] } = useTopicConfig(props); + + const columns = React.useMemo<ColumnDef<TopicConfig>[]>( + () => [ + { + header: 'Key', + accessorKey: 'name', + cell: ValueCell, + }, + { + header: 'Value', + accessorKey: 'value', + cell: ValueCell, + }, + { + header: 'Default Value', + accessorKey: 'defaultValue', + cell: DefaultValueCell, + }, + ], + [] ); + + return <Table columns={columns} data={data} />; }; export default Settings; diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Settings/__test__/ConfigListItem.spec.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Settings/__test__/ConfigListItem.spec.tsx deleted file mode 100644 index b363ff55337..00000000000 --- a/kafka-ui-react-app/src/components/Topics/Topic/Settings/__test__/ConfigListItem.spec.tsx +++ /dev/null @@ -1,45 +0,0 @@ -import React from 'react'; -import { render } from 'lib/testHelpers'; -import { screen } from '@testing-library/react'; -import ConfigListItem, { - ListItemProps, -} from 'components/Topics/Topic/Settings/ConfigListItem'; - -const setupComponent = (props: ListItemProps) => { - render( - <table> - <tbody> - <ConfigListItem {...props} /> - </tbody> - </table> - ); -}; - -const getName = () => screen.getByText('someName'); -const getValue = () => screen.getByText('someValue'); - -it('renders with CustomValue', () => { - setupComponent({ - config: { - name: 'someName', - value: 'someValue', - defaultValue: 'someDefaultValue', - }, - }); - expect(getName()).toBeInTheDocument(); - expect(getName()).toHaveStyle('font-weight: 500'); - expect(getValue()).toBeInTheDocument(); - expect(getValue()).toHaveStyle('font-weight: 500'); - expect(screen.getByText('someDefaultValue')).toBeInTheDocument(); -}); - -it('renders without CustomValue', () => { - setupComponent({ - config: { name: 'someName', value: 'someValue', defaultValue: 'someValue' }, - }); - expect(getName()).toBeInTheDocument(); - expect(getName()).toHaveStyle('font-weight: 400'); - expect(getValue()).toBeInTheDocument(); - expect(getValue()).toHaveStyle('font-weight: 400'); - expect(screen.getByTitle('Default Value')).toHaveTextContent(''); -}); diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Settings/__test__/Settings.spec.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Settings/__test__/Settings.spec.tsx index cf5236ffa15..c2aa9a93e5d 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Settings/__test__/Settings.spec.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Settings/__test__/Settings.spec.tsx @@ -13,11 +13,8 @@ jest.mock('lib/hooks/api/topics', () => ({ useTopicConfig: jest.fn(), })); -jest.mock('components/Topics/Topic/Settings/ConfigListItem', () => () => ( - <tr> - <td>ConfigListItemMock</td> - </tr> -)); +const getName = () => screen.getByText('compression.type'); +const getValue = () => screen.getByText('producer'); describe('Settings', () => { const renderComponent = () => { @@ -37,9 +34,10 @@ describe('Settings', () => { renderComponent(); }); - it('should check it returns null if no config is passed', () => { - expect(screen.getByRole('table')).toBeInTheDocument(); - const items = screen.getAllByText('ConfigListItemMock'); - expect(items.length).toEqual(topicConfigPayload.length); + it('renders without CustomValue', () => { + expect(getName()).toBeInTheDocument(); + expect(getName()).toHaveStyle('font-weight: 400'); + expect(getValue()).toBeInTheDocument(); + expect(getValue()).toHaveStyle('font-weight: 400'); }); });
null
test
val
2022-11-09T12:52:20
"2022-09-30T10:57:44Z"
workshur
train
provectus/kafka-ui/2725_2835
provectus/kafka-ui
provectus/kafka-ui/2725
provectus/kafka-ui/2835
[ "connected" ]
2853b34b5f4667e22dac5932a5000e740315bdb8
d2a62e27628cb55b88258df6a1f13b6faccb651d
[ "hey @Haarolean assign me this issue", "@yogesh-9999 any luck?", "@Haarolean can I take that issue?", "@vitalymatyushik yeah it seems" ]
[]
"2022-10-27T20:33:20Z"
[ "type/enhancement", "good first issue", "scope/frontend", "status/accepted", "hacktoberfest" ]
Make topic bulk actions buttons always visible
<img width="747" alt="image" src="https://user-images.githubusercontent.com/1494347/195046352-d2b27d74-bbef-46a2-982c-c2aece1cb7df.png"> 1. Make these buttons visible even if no topics are selected 2. Make the buttons disabled if no topics are selected
[ "kafka-ui-react-app/src/components/Topics/List/BatchActionsBar.tsx", "kafka-ui-react-app/src/components/common/NewTable/Table.tsx", "kafka-ui-react-app/src/components/common/NewTable/__test__/Table.spec.tsx" ]
[ "kafka-ui-react-app/src/components/Topics/List/BatchActionsBar.tsx", "kafka-ui-react-app/src/components/common/NewTable/Table.tsx", "kafka-ui-react-app/src/components/common/NewTable/__test__/Table.spec.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/List/BatchActionsBar.tsx b/kafka-ui-react-app/src/components/Topics/List/BatchActionsBar.tsx index a5acec28f5b..3ff45cfed78 100644 --- a/kafka-ui-react-app/src/components/Topics/List/BatchActionsBar.tsx +++ b/kafka-ui-react-app/src/components/Topics/List/BatchActionsBar.tsx @@ -63,6 +63,12 @@ const BatchActionsbar: React.FC<BatchActionsbarProps> = ({ type Tuple = [string, string]; const getCopyTopicPath = () => { + if (!rows.length) { + return { + pathname: '', + search: '', + }; + } const topic = rows[0].original; const search = Object.keys(topic).reduce((acc: Tuple[], key) => { @@ -86,13 +92,14 @@ const BatchActionsbar: React.FC<BatchActionsbarProps> = ({ buttonSize="M" buttonType="secondary" onClick={deleteTopicsHandler} + disabled={!selectedTopics.length} > Delete selected topics </Button> <Button buttonSize="M" buttonType="secondary" - disabled={selectedTopics.length > 1} + disabled={selectedTopics.length !== 1} to={getCopyTopicPath()} > Copy selected topic @@ -101,6 +108,7 @@ const BatchActionsbar: React.FC<BatchActionsbarProps> = ({ buttonSize="M" buttonType="secondary" onClick={purgeTopicsHandler} + disabled={!selectedTopics.length} > Purge messages of selected topics </Button> diff --git a/kafka-ui-react-app/src/components/common/NewTable/Table.tsx b/kafka-ui-react-app/src/components/common/NewTable/Table.tsx index 7e3b571c375..2c80ac18b27 100644 --- a/kafka-ui-react-app/src/components/common/NewTable/Table.tsx +++ b/kafka-ui-react-app/src/components/common/NewTable/Table.tsx @@ -101,7 +101,7 @@ const getSortingFromSearchParams = (searchParams: URLSearchParams) => { * - use `enableRowSelection` prop to enable row selection. This prop can be a boolean or * a function that returns true if the particular row can be selected. * - use `batchActionsBar` prop to provide a component that will be rendered at the top of the table - * when row selection is enabled and there are selected rows. + * when row selection is enabled. * * 5. Server side processing: * - set `serverSideProcessing` to true @@ -190,7 +190,7 @@ const Table: React.FC<TableProps<any>> = ({ return ( <> - {table.getSelectedRowModel().flatRows.length > 0 && BatchActionsBar && ( + {BatchActionsBar && ( <S.TableActionsBar> <BatchActionsBar rows={table.getSelectedRowModel().flatRows} diff --git a/kafka-ui-react-app/src/components/common/NewTable/__test__/Table.spec.tsx b/kafka-ui-react-app/src/components/common/NewTable/__test__/Table.spec.tsx index 227dc552059..dbfdddab662 100644 --- a/kafka-ui-react-app/src/components/common/NewTable/__test__/Table.spec.tsx +++ b/kafka-ui-react-app/src/components/common/NewTable/__test__/Table.spec.tsx @@ -318,7 +318,7 @@ describe('Table', () => { it('renders action bar', async () => { expect(screen.getAllByRole('row').length).toEqual(data.length + 1); - expect(screen.queryByText('I am Action Bar')).not.toBeInTheDocument(); + expect(screen.queryByText('I am Action Bar')).toBeInTheDocument(); const checkboxes = screen.getAllByRole('checkbox'); expect(checkboxes.length).toEqual(data.length + 1); await userEvent.click(checkboxes[2]);
null
train
val
2022-10-26T22:03:42
"2022-10-11T09:00:10Z"
Haarolean
train
provectus/kafka-ui/2840_2841
provectus/kafka-ui
provectus/kafka-ui/2840
provectus/kafka-ui/2841
[ "connected" ]
95e494ea09996c673ba2a876f5b829c066a2d9b4
b28492acd4a84f12bfcf8987f03c9e41c5410e65
[]
[]
"2022-10-28T14:00:06Z"
[ "scope/frontend", "status/accepted", "type/chore" ]
Wrong type declaration in the Brokers Fixtures file for the value metrics
<!-- Don't forget to check for existing issues/discussions regarding your proposal. We might already have it. https://github.com/provectus/kafka-ui/issues https://github.com/provectus/kafka-ui/discussions --> **Describe the bug** <!--(A clear and concise description of what the bug is.)--> Wrong type declaration in the Brokers for the metrics value fixtures file , which leads the file to be read as an invalid one by the editor **Set up** <!-- How do you run the app? Please provide as much info as possible: 1. App version (docker image version or check commit hash in the top left corner in UI) 2. Helm chart version, if you use one 3. Any IAAC configs We might close the issue without further explanation if you don't provide such information. --> Latest version from the master branch. **Steps to Reproduce** <!-- We'd like you to provide an example setup (via docker-compose, helm, etc.) to reproduce the problem, especially with a complex setups. --> Steps to reproduce the behavior: 1. Go to the src/components/Brokers/utils/__tests__/fixtures.ts and wait and that file will be automatically picked up as a invalid type. **Expected behavior** <!-- (A clear and concise description of what you expected to happen) --> value variable should have the correct type which is number or undefined. **Screenshots** <!-- (If applicable, add screenshots to help explain your problem) --> <img width="1035" alt="image" src="https://user-images.githubusercontent.com/46796009/198612331-729e4e86-5670-44df-8ffd-2fc6bcf223f4.png"> <img width="341" alt="image" src="https://user-images.githubusercontent.com/46796009/198612495-ba141e30-ac9f-4388-a49a-9e45c7ad3849.png">
[ "kafka-ui-react-app/src/components/Brokers/utils/__test__/fixtures.ts" ]
[ "kafka-ui-react-app/src/components/Brokers/utils/__test__/fixtures.ts" ]
[]
diff --git a/kafka-ui-react-app/src/components/Brokers/utils/__test__/fixtures.ts b/kafka-ui-react-app/src/components/Brokers/utils/__test__/fixtures.ts index 9ab3e4b662d..310c9bc54a3 100644 --- a/kafka-ui-react-app/src/components/Brokers/utils/__test__/fixtures.ts +++ b/kafka-ui-react-app/src/components/Brokers/utils/__test__/fixtures.ts @@ -6,169 +6,20 @@ export const brokerMetricsPayload: BrokerMetrics = { metrics: [ { name: 'TotalFetchRequestsPerSec', - canonicalName: - 'kafka.server:name=TotalFetchRequestsPerSec,topic=_connect_status,type=BrokerTopicMetrics', - params: { - topic: '_connect_status', - name: 'TotalFetchRequestsPerSec', - type: 'BrokerTopicMetrics', - }, - value: { - OneMinuteRate: 19.408369293127542, - FifteenMinuteRate: 19.44631556589501, - Count: 191615, - FiveMinuteRate: 19.464393718807774, - MeanRate: 19.4233855043407, + labels: { + canonicalName: + 'kafka.server:name=TotalFetchRequestsPerSec,topic=_connect_status,type=BrokerTopicMetrics', }, + value: 10, }, { name: 'ZooKeeperRequestLatencyMs', - canonicalName: - 'kafka.server:name=ZooKeeperRequestLatencyMs,type=ZooKeeperClientMetrics', - params: { - name: 'ZooKeeperRequestLatencyMs', - type: 'ZooKeeperClientMetrics', - }, - value: { - Mean: 4.907351022183558, - StdDev: 10.589608223906348, - '75thPercentile': 2, - '98thPercentile': 10, - Min: 0, - '95thPercentile': 5, - '99thPercentile': 15, - Max: 151, - '999thPercentile': 92.79700000000003, - Count: 2301, - '50thPercentile': 1, - }, + value: 11, }, { name: 'RequestHandlerAvgIdlePercent', - canonicalName: - 'kafka.server:name=RequestHandlerAvgIdlePercent,type=KafkaRequestHandlerPool', - params: { - name: 'RequestHandlerAvgIdlePercent', - type: 'KafkaRequestHandlerPool', - }, - value: { - OneMinuteRate: 0.9999008788765713, - FifteenMinuteRate: 0.9983845959639047, - Count: 9937344680371, - FiveMinuteRate: 0.9986337207880311, - MeanRate: 0.9971616923696525, - }, - }, - { - name: 'BytesInPerSec', - canonicalName: - 'kafka.server:name=BytesInPerSec,topic=_connect_status,type=BrokerTopicMetrics', - params: { - topic: '_connect_status', - name: 'BytesInPerSec', - type: 'BrokerTopicMetrics', - }, - value: { - OneMinuteRate: 0, - FifteenMinuteRate: 0, - Count: 0, - FiveMinuteRate: 0, - MeanRate: 0, - }, - }, - { - name: 'FetchMessageConversionsPerSec', - canonicalName: - 'kafka.server:name=FetchMessageConversionsPerSec,topic=__consumer_offsets,type=BrokerTopicMetrics', - params: { - topic: '__consumer_offsets', - name: 'FetchMessageConversionsPerSec', - type: 'BrokerTopicMetrics', - }, - value: { - OneMinuteRate: 0, - FifteenMinuteRate: 0, - Count: 0, - FiveMinuteRate: 0, - MeanRate: 0, - }, - }, - { - name: 'TotalProduceRequestsPerSec', - canonicalName: - 'kafka.server:name=TotalProduceRequestsPerSec,topic=_connect_status,type=BrokerTopicMetrics', - params: { - topic: '_connect_status', - name: 'TotalProduceRequestsPerSec', - type: 'BrokerTopicMetrics', - }, - value: { - OneMinuteRate: 0, - FifteenMinuteRate: 0, - Count: 0, - FiveMinuteRate: 0, - MeanRate: 0, - }, - }, - { - name: 'MaxLag', - canonicalName: - 'kafka.server:clientId=Replica,name=MaxLag,type=ReplicaFetcherManager', - params: { - clientId: 'Replica', - name: 'MaxLag', - type: 'ReplicaFetcherManager', - }, - value: { - Value: 0, - }, - }, - { - name: 'UnderMinIsrPartitionCount', - canonicalName: - 'kafka.server:name=UnderMinIsrPartitionCount,type=ReplicaManager', - params: { - name: 'UnderMinIsrPartitionCount', - type: 'ReplicaManager', - }, - value: { - Value: 0, - }, - }, - { - name: 'ZooKeeperDisconnectsPerSec', - canonicalName: - 'kafka.server:name=ZooKeeperDisconnectsPerSec,type=SessionExpireListener', - params: { - name: 'ZooKeeperDisconnectsPerSec', - type: 'SessionExpireListener', - }, - value: { - OneMinuteRate: 0, - FifteenMinuteRate: 0, - Count: 0, - FiveMinuteRate: 0, - MeanRate: 0, - }, - }, - { - name: 'BytesInPerSec', - canonicalName: - 'kafka.server:name=BytesInPerSec,topic=__confluent.support.metrics,type=BrokerTopicMetrics', - params: { - topic: '__confluent.support.metrics', - name: 'BytesInPerSec', - type: 'BrokerTopicMetrics', - }, - value: { - OneMinuteRate: 3.093893673470914e-70, - FifteenMinuteRate: 0.004057932469784932, - Count: 1263, - FiveMinuteRate: 1.047243693828501e-12, - MeanRate: 0.12704831069266603, - }, }, ], }; export const transformedBrokerMetricsPayload = - '{"segmentSize":23,"segmentCount":23,"metrics":[{"name":"TotalFetchRequestsPerSec","canonicalName":"kafka.server:name=TotalFetchRequestsPerSec,topic=_connect_status,type=BrokerTopicMetrics","params":{"topic":"_connect_status","name":"TotalFetchRequestsPerSec","type":"BrokerTopicMetrics"},"value":{"OneMinuteRate":19.408369293127542,"FifteenMinuteRate":19.44631556589501,"Count":191615,"FiveMinuteRate":19.464393718807774,"MeanRate":19.4233855043407}},{"name":"ZooKeeperRequestLatencyMs","canonicalName":"kafka.server:name=ZooKeeperRequestLatencyMs,type=ZooKeeperClientMetrics","params":{"name":"ZooKeeperRequestLatencyMs","type":"ZooKeeperClientMetrics"},"value":{"Mean":4.907351022183558,"StdDev":10.589608223906348,"75thPercentile":2,"98thPercentile":10,"Min":0,"95thPercentile":5,"99thPercentile":15,"Max":151,"999thPercentile":92.79700000000003,"Count":2301,"50thPercentile":1}},{"name":"RequestHandlerAvgIdlePercent","canonicalName":"kafka.server:name=RequestHandlerAvgIdlePercent,type=KafkaRequestHandlerPool","params":{"name":"RequestHandlerAvgIdlePercent","type":"KafkaRequestHandlerPool"},"value":{"OneMinuteRate":0.9999008788765713,"FifteenMinuteRate":0.9983845959639047,"Count":9937344680371,"FiveMinuteRate":0.9986337207880311,"MeanRate":0.9971616923696525}},{"name":"BytesInPerSec","canonicalName":"kafka.server:name=BytesInPerSec,topic=_connect_status,type=BrokerTopicMetrics","params":{"topic":"_connect_status","name":"BytesInPerSec","type":"BrokerTopicMetrics"},"value":{"OneMinuteRate":0,"FifteenMinuteRate":0,"Count":0,"FiveMinuteRate":0,"MeanRate":0}},{"name":"FetchMessageConversionsPerSec","canonicalName":"kafka.server:name=FetchMessageConversionsPerSec,topic=__consumer_offsets,type=BrokerTopicMetrics","params":{"topic":"__consumer_offsets","name":"FetchMessageConversionsPerSec","type":"BrokerTopicMetrics"},"value":{"OneMinuteRate":0,"FifteenMinuteRate":0,"Count":0,"FiveMinuteRate":0,"MeanRate":0}},{"name":"TotalProduceRequestsPerSec","canonicalName":"kafka.server:name=TotalProduceRequestsPerSec,topic=_connect_status,type=BrokerTopicMetrics","params":{"topic":"_connect_status","name":"TotalProduceRequestsPerSec","type":"BrokerTopicMetrics"},"value":{"OneMinuteRate":0,"FifteenMinuteRate":0,"Count":0,"FiveMinuteRate":0,"MeanRate":0}},{"name":"MaxLag","canonicalName":"kafka.server:clientId=Replica,name=MaxLag,type=ReplicaFetcherManager","params":{"clientId":"Replica","name":"MaxLag","type":"ReplicaFetcherManager"},"value":{"Value":0}},{"name":"UnderMinIsrPartitionCount","canonicalName":"kafka.server:name=UnderMinIsrPartitionCount,type=ReplicaManager","params":{"name":"UnderMinIsrPartitionCount","type":"ReplicaManager"},"value":{"Value":0}},{"name":"ZooKeeperDisconnectsPerSec","canonicalName":"kafka.server:name=ZooKeeperDisconnectsPerSec,type=SessionExpireListener","params":{"name":"ZooKeeperDisconnectsPerSec","type":"SessionExpireListener"},"value":{"OneMinuteRate":0,"FifteenMinuteRate":0,"Count":0,"FiveMinuteRate":0,"MeanRate":0}},{"name":"BytesInPerSec","canonicalName":"kafka.server:name=BytesInPerSec,topic=__confluent.support.metrics,type=BrokerTopicMetrics","params":{"topic":"__confluent.support.metrics","name":"BytesInPerSec","type":"BrokerTopicMetrics"},"value":{"OneMinuteRate":3.093893673470914e-70,"FifteenMinuteRate":0.004057932469784932,"Count":1263,"FiveMinuteRate":1.047243693828501e-12,"MeanRate":0.12704831069266603}}]}'; + '{"segmentSize":23,"segmentCount":23,"metrics":[{"name":"TotalFetchRequestsPerSec","labels":{"canonicalName":"kafka.server:name=TotalFetchRequestsPerSec,topic=_connect_status,type=BrokerTopicMetrics"},"value":10},{"name":"ZooKeeperRequestLatencyMs","value":11},{"name":"RequestHandlerAvgIdlePercent"}]}';
null
test
val
2022-10-28T18:44:01
"2022-10-28T13:33:14Z"
Mgrdich
train
provectus/kafka-ui/2829_2842
provectus/kafka-ui
provectus/kafka-ui/2829
provectus/kafka-ui/2842
[ "connected" ]
d2a62e27628cb55b88258df6a1f13b6faccb651d
95e494ea09996c673ba2a876f5b829c066a2d9b4
[ "Hello there treyhendon! πŸ‘‹\n\nThank you and congratulations πŸŽ‰ for opening your very first issue in this project! πŸ’–\n\nIn case you want to claim this issue, please comment down below! We will try to get back to you as soon as we can. πŸ‘€", "Done in #2842, thank you" ]
[]
"2022-10-28T16:37:32Z"
[ "status/accepted", "scope/k8s" ]
0.4.0 Image Tag Does Not Match
**Steps to Reproduce** Steps to reproduce the behavior: 1. Deploy current main helm chart **Expected behavior** The chart defaults that if no tag is defined, the chart app version is used. Published Docker Tag: `v0.4.0` Helm Helper: ``` {{- $tag := .Values.image.tag | default .Chart.AppVersion -}} ``` Chart Value: ``` appVersion: 0.4.0 ``` **Screenshots** Docker: ![image](https://user-images.githubusercontent.com/3513589/198065920-feb93aef-254c-49d4-851d-e2c800a0a4a0.png) Deployed Tag: <img width="356" alt="image" src="https://user-images.githubusercontent.com/3513589/198065992-6d6370ce-f14b-43a3-bd25-cfb68280f858.png"> **Additional context** Not sure if you'd rather update the appVersion in chart to be `"v0.4.0"` or publish a new tag to Docker.
[ "charts/kafka-ui/Chart.yaml" ]
[ "charts/kafka-ui/Chart.yaml" ]
[]
diff --git a/charts/kafka-ui/Chart.yaml b/charts/kafka-ui/Chart.yaml index 89982674019..2da8388a44a 100644 --- a/charts/kafka-ui/Chart.yaml +++ b/charts/kafka-ui/Chart.yaml @@ -3,5 +3,5 @@ name: kafka-ui description: A Helm chart for kafka-UI type: application version: 0.4.4 -appVersion: 0.4.0 +appVersion: v0.4.0 icon: https://github.com/provectus/kafka-ui/raw/master/documentation/images/kafka-ui-logo.png
null
train
val
2022-10-28T17:55:35
"2022-10-26T15:20:39Z"
treyhendon
train
provectus/kafka-ui/2838_2843
provectus/kafka-ui
provectus/kafka-ui/2838
provectus/kafka-ui/2843
[ "connected" ]
b28492acd4a84f12bfcf8987f03c9e41c5410e65
81072541a8aab5271c63ae168f44e3ac86fcce16
[]
[]
"2022-10-29T10:16:31Z"
[ "type/documentation", "scope/frontend" ]
React Documentation Readme Includes a Link to create-react-app
I don't think create react app link is relevant in the documentation , since nowadays the app uses Vite as a module bundler , where create-react-app uses Webpack and we don't rely on react-script to build the application, we should replace it with Vite source code link. <img width="1712" alt="image" src="https://user-images.githubusercontent.com/46796009/198550641-1b383bbd-dda5-4927-bd8f-e1979529a160.png">
[ "kafka-ui-react-app/README.md" ]
[ "kafka-ui-react-app/README.md" ]
[]
diff --git a/kafka-ui-react-app/README.md b/kafka-ui-react-app/README.md index 8e5ea6ebf3b..8ab814cd1d5 100644 --- a/kafka-ui-react-app/README.md +++ b/kafka-ui-react-app/README.md @@ -66,4 +66,4 @@ pnpm start ``` ## Links -* [Create React App](https://github.com/facebook/create-react-app) +* [Vite](https://github.com/vitejs/vite)
null
train
val
2022-10-29T15:21:37
"2022-10-28T09:21:41Z"
Mgrdich
train
provectus/kafka-ui/2759_2846
provectus/kafka-ui
provectus/kafka-ui/2759
provectus/kafka-ui/2846
[ "connected" ]
81072541a8aab5271c63ae168f44e3ac86fcce16
2f2e19d1bce41ec8c0f867b20b4c72b20f764bf0
[ "The internality flag is being defaulted to a single underscore character now. Can be overridden with \"kafka_clusters_internalTopicPrefix\" property." ]
[]
"2022-10-30T15:39:25Z"
[ "type/bug", "good first issue", "scope/backend", "status/accepted", "status/confirmed", "hacktoberfest" ]
Unable to delete topics with names beginning with an underscore
### Discussed in https://github.com/provectus/kafka-ui/discussions/2755 <div type='discussions-op-text'> <sup>Originally posted by **nitin1677** October 14, 2022</sup> Kafka UI Version: d149d26013a81e22e5d0c8b59db56ac0565b9e55 ( [d149d26](https://github.com/provectus/kafka-ui/commit/d149d26)) Remove topic options are disabled for topics, where topic-name starts with underscore (_). what are the options to enable delete for topics, where name starts with underscore (_).</div>
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/InternalTopic.java" ]
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/InternalTopic.java" ]
[]
diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/InternalTopic.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/InternalTopic.java index b669d0db414..a553165de4d 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/InternalTopic.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/model/InternalTopic.java @@ -42,9 +42,7 @@ public static InternalTopic from(TopicDescription topicDescription, Metrics metrics, InternalLogDirStats logDirInfo) { var topic = InternalTopic.builder(); - topic.internal( - topicDescription.isInternal() || topicDescription.name().startsWith("_") - ); + topic.internal(topicDescription.isInternal()); topic.name(topicDescription.name()); List<InternalPartition> partitions = topicDescription.partitions().stream()
null
train
val
2022-10-29T20:45:21
"2022-10-17T09:12:49Z"
Haarolean
train
provectus/kafka-ui/2716_2847
provectus/kafka-ui
provectus/kafka-ui/2716
provectus/kafka-ui/2847
[ "connected" ]
81072541a8aab5271c63ae168f44e3ac86fcce16
cb8049e769452a40007c29acdfae43e5e8abe53d
[ "Hey, thanks for reaching out. We're aware, the fix will be available in a few days. Meanwhile, you can set the env variable `timestamp.format` to `DD.MM.YYYY HH:mm:ss`.", "thx, adding environment variable for timestamp.format helped to correct the date format.", "> Hey, thanks for reaching out. We're aware, the fix will be available in a few days. Meanwhile, you can set the env variable `timestamp.format` to `DD.MM.YYYY HH:mm:ss`.\r\n\r\n@Haarolean @nitin1677 I don't understand how to do this, can you write in steps? what is the name of the variable itself and how to export it, this is done through the UI or inside the docker container or at startup\r\nI tried docker exec\r\ntimestamp.format=DD.MM.YYYY HH:mm:ss\r\ntimestamp_format=DD.MM.YYYY HH:mm:ss\r\nTIMESTAMP_FORMAT=DD.MM.YYYY HH:mm:ss\r\nTIMESTAMP_FORMAT=\"DD.MM.YYYY HH:mm:ss\"" ]
[]
"2022-10-30T15:55:58Z"
[ "type/bug", "scope/backend", "scope/frontend", "status/accepted", "status/confirmed" ]
Invalid dates after a recent pull
<!-- Don't forget to check for existing issues/discussions regarding your proposal. We might already have it. https://github.com/provectus/kafka-ui/issues https://github.com/provectus/kafka-ui/discussions --> **Describe the bug** <!--(A clear and concise description of what the bug is.)--> After i pulled the recent image from master branch, the dates are badly messed up and does not show in correct/readable format. for e.g. top left, current commit shows like this, notice the date on left side before current commit in brackes. Fr.10.2022 11:20:53 (7d5b7de) Also from message viewer, the timestamp formats also messed up with format. Offset Partition Timestamp Key 10517 11 Fr.09.2022 17:30:21 238087980 **Set up** <!-- How do you run the app? Please provide as much info as possible: 1. App version (docker image version or check commit hash in the top left corner in UI) 2. Helm chart version, if you use one 3. Any IAAC configs We might close the issue without further explanation if you don't provide such information. --> Commit Hash from top left corner in UI: Fr.10.2022 11:20:53 (7d5b7de) **Steps to Reproduce** <!-- We'd like you to provide an example setup (via docker-compose, helm, etc.) to reproduce the problem, especially with a complex setups. --> Steps to reproduce the behavior: 1. docker pull provectuslabs/kafka-ui:master **Expected behavior** <!-- (A clear and concise description of what you expected to happen) --> Results from Commit hash: d149d26013a81e22e5d0c8b59db56ac0565b9e55 ( [d149d26](https://github.com/provectus/kafka-ui/commit/d149d26) ) Message Viewer: Offset|Partition| Timestamp| Key 10389 | 1 | 09.23.22 07:39:30 pm | 238087985 **Screenshots** <!-- (If applicable, add screenshots to help explain your problem) --> **Additional context** <!-- (Add any other context about the problem here) -->
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/controller/InfoController.java", "kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml", "kafka-ui-react-app/src/lib/hooks/api/__tests__/timeFormat.spec.ts", "kafka-ui-react-app/src/lib/hooks/api/timeFormat.ts" ]
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/controller/InfoController.java", "kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml", "kafka-ui-react-app/src/lib/hooks/api/__tests__/timeFormat.spec.ts", "kafka-ui-react-app/src/lib/hooks/api/timeFormat.ts" ]
[]
diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/controller/InfoController.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/controller/InfoController.java index 66e5d70bd33..cdda3d0953e 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/controller/InfoController.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/controller/InfoController.java @@ -17,9 +17,16 @@ public class InfoController extends AbstractController implements TimeStampForma @Value("${timestamp.format:dd.MM.YYYY HH:mm:ss}") private String timeStampFormat; + @Value("${timestamp.format:DD.MM.YYYY HH:mm:ss}") + private String timeStampFormatIso; @Override public Mono<ResponseEntity<TimeStampFormatDTO>> getTimeStampFormat(ServerWebExchange exchange) { return Mono.just(ResponseEntity.ok(new TimeStampFormatDTO().timeStampFormat(timeStampFormat))); } + + @Override + public Mono<ResponseEntity<TimeStampFormatDTO>> getTimeStampFormatISO(ServerWebExchange exchange) { + return Mono.just(ResponseEntity.ok(new TimeStampFormatDTO().timeStampFormat(timeStampFormatIso))); + } } diff --git a/kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml b/kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml index c4765fb2202..5ea470b4f95 100644 --- a/kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml +++ b/kafka-ui-contract/src/main/resources/swagger/kafka-ui-api.yaml @@ -1733,7 +1733,7 @@ paths: get: tags: - TimeStampFormat - summary: getTimeStampFormat + summary: get system default datetime format operationId: getTimeStampFormat responses: 200: @@ -1742,6 +1742,21 @@ paths: application/json: schema: $ref: '#/components/schemas/TimeStampFormat' + + /api/info/timestampformat/iso: + get: + tags: + - TimeStampFormat + summary: get system default datetime format (in ISO format, for JS) + operationId: getTimeStampFormatISO + responses: + 200: + description: OK + content: + application/json: + schema: + $ref: '#/components/schemas/TimeStampFormat' + components: schemas: TopicSerdeSuggestion: diff --git a/kafka-ui-react-app/src/lib/hooks/api/__tests__/timeFormat.spec.ts b/kafka-ui-react-app/src/lib/hooks/api/__tests__/timeFormat.spec.ts index cc080f58e14..c6984c40b15 100644 --- a/kafka-ui-react-app/src/lib/hooks/api/__tests__/timeFormat.spec.ts +++ b/kafka-ui-react-app/src/lib/hooks/api/__tests__/timeFormat.spec.ts @@ -3,7 +3,7 @@ import { expectQueryWorks, renderQueryHook } from 'lib/testHelpers'; import * as hooks from 'lib/hooks/api/timeFormat'; import { timeFormatPayload } from 'lib/fixtures/timeFormat'; -const timeFormatPath = '/api/info/timestampformat'; +const timeFormatPath = '/api/info/timestampformat/iso'; describe('Time format hooks', () => { beforeEach(() => fetchMock.restore()); diff --git a/kafka-ui-react-app/src/lib/hooks/api/timeFormat.ts b/kafka-ui-react-app/src/lib/hooks/api/timeFormat.ts index fcbdf931520..3b36f74066f 100644 --- a/kafka-ui-react-app/src/lib/hooks/api/timeFormat.ts +++ b/kafka-ui-react-app/src/lib/hooks/api/timeFormat.ts @@ -5,7 +5,7 @@ import { QUERY_REFETCH_OFF_OPTIONS } from 'lib/constants'; export function useTimeFormat() { return useQuery( ['settings', 'timestampformat'], - () => api.getTimeStampFormat(), + () => api.getTimeStampFormatISO(), QUERY_REFETCH_OFF_OPTIONS ); }
null
train
val
2022-10-29T20:45:21
"2022-10-08T22:48:06Z"
nitin1677
train
provectus/kafka-ui/2676_2848
provectus/kafka-ui
provectus/kafka-ui/2676
provectus/kafka-ui/2848
[ "connected" ]
1b1caddd310d3c92c8f3430b991cc48efef469ff
b0782da7d137fba481eb9059a60e052617255b2a
[ "i wanna work on this, Can you please tell me any easy way of finding all the sizes in the application.\r\nthanks", "@haseeb-xd sure, go ahead. \r\nThey might be formatted via the same function through the app. Topic statistics is one of the places with a lot of sizes.", "@haseeb-xd any luck?" ]
[]
"2022-10-30T19:11:24Z"
[ "good first issue", "scope/frontend", "status/accepted", "type/chore" ]
Add a space to all formatted sizes
All sizes in the app are displayed like this: "123KB". Please add a space between the numeric and size parts.
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java", "kafka-ui-react-app/src/components/Brokers/Broker/__test__/Broker.spec.tsx", "kafka-ui-react-app/src/components/Topics/List/__tests__/TopicTable.spec.tsx", "kafka-ui-react-app/src/components/common/BytesFormatted/BytesFormatted.tsx", "kafka-ui-react-app/src/components/common/BytesFormatted/__tests__/BytesFormatted.spec.tsx", "kafka-ui-react-app/src/components/common/NewTable/__test__/Table.spec.tsx" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java", "kafka-ui-react-app/src/components/Brokers/Broker/__test__/Broker.spec.tsx", "kafka-ui-react-app/src/components/Topics/List/__tests__/TopicTable.spec.tsx", "kafka-ui-react-app/src/components/common/BytesFormatted/BytesFormatted.styled.ts", "kafka-ui-react-app/src/components/common/BytesFormatted/BytesFormatted.tsx", "kafka-ui-react-app/src/components/common/BytesFormatted/__tests__/BytesFormatted.spec.tsx", "kafka-ui-react-app/src/components/common/NewTable/__test__/Table.spec.tsx" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java index e619dc5c95d..d0632817b7d 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java @@ -21,7 +21,7 @@ public class TopicDetails { protected SelenideElement clearMessagesBtn = $x(("//div[contains(text(), 'Clear messages')]")); protected SelenideElement overviewTab = $x("//a[contains(text(),'Overview')]"); protected SelenideElement messagesTab = $x("//a[contains(text(),'Messages')]"); - protected SelenideElement editSettingsTab = $x("//li[@role][contains(text(),'Edit settings')]"); + protected SelenideElement editSettingsMenu = $x("//li[@role][contains(text(),'Edit settings')]"); protected SelenideElement removeTopicBtn = $x("//ul[@role='menu']//div[contains(text(),'Remove Topic')]"); protected SelenideElement confirmBtn = $x("//div[@role='dialog']//button[contains(text(),'Confirm')]"); protected SelenideElement produceMessageBtn = $x("//div//button[text()='Produce Message']"); @@ -39,7 +39,7 @@ public TopicDetails waitUntilScreenReady() { } @Step - public TopicDetails openTopicMenu(TopicMenu menu) { + public TopicDetails openDetailsTab(TopicMenu menu) { $(By.linkText(menu.getValue())).shouldBe(Condition.visible).click(); return this; } @@ -52,7 +52,7 @@ public TopicDetails openDotMenu() { @Step public TopicDetails clickEditSettingsMenu() { - editSettingsTab.shouldBe(Condition.visible).click(); + editSettingsMenu.shouldBe(Condition.visible).click(); return this; } diff --git a/kafka-ui-react-app/src/components/Brokers/Broker/__test__/Broker.spec.tsx b/kafka-ui-react-app/src/components/Brokers/Broker/__test__/Broker.spec.tsx index 9723149828c..ae6a7f4b3b6 100644 --- a/kafka-ui-react-app/src/components/Brokers/Broker/__test__/Broker.spec.tsx +++ b/kafka-ui-react-app/src/components/Brokers/Broker/__test__/Broker.spec.tsx @@ -66,7 +66,7 @@ describe('Broker Component', () => { expect( screen.getByText(brokerDiskUsage?.segmentCount || '') ).toBeInTheDocument(); - expect(screen.getByText('12MB')).toBeInTheDocument(); + expect(screen.getByText('12 MB')).toBeInTheDocument(); expect(screen.getByText('Segment Count')).toBeInTheDocument(); expect( diff --git a/kafka-ui-react-app/src/components/Topics/List/__tests__/TopicTable.spec.tsx b/kafka-ui-react-app/src/components/Topics/List/__tests__/TopicTable.spec.tsx index 4b5f046cd9d..833b0825184 100644 --- a/kafka-ui-react-app/src/components/Topics/List/__tests__/TopicTable.spec.tsx +++ b/kafka-ui-react-app/src/components/Topics/List/__tests__/TopicTable.spec.tsx @@ -92,13 +92,13 @@ describe('TopicTable Components', () => { screen.getByRole('link', { name: '__internal.topic' }) ).toBeInTheDocument(); expect( - screen.getByRole('row', { name: '__internal.topic 1 0 1 0 0Bytes' }) + screen.getByRole('row', { name: '__internal.topic 1 0 1 0 0 Bytes' }) ).toBeInTheDocument(); expect( screen.getByRole('link', { name: 'external.topic' }) ).toBeInTheDocument(); expect( - screen.getByRole('row', { name: 'external.topic 1 0 1 0 1KB' }) + screen.getByRole('row', { name: 'external.topic 1 0 1 0 1 KB' }) ).toBeInTheDocument(); expect(screen.getAllByRole('checkbox').length).toEqual(3); @@ -211,7 +211,7 @@ describe('TopicTable Components', () => { ).toEqual(2); // Internal topic action buttons are disabled const internalTopicRow = screen.getByRole('row', { - name: '__internal.topic 1 0 1 0 0Bytes', + name: '__internal.topic 1 0 1 0 0 Bytes', }); expect(internalTopicRow).toBeInTheDocument(); expect( @@ -221,7 +221,7 @@ describe('TopicTable Components', () => { ).toBeDisabled(); // External topic action buttons are enabled const externalTopicRow = screen.getByRole('row', { - name: 'external.topic 1 0 1 0 1KB', + name: 'external.topic 1 0 1 0 1 KB', }); expect(externalTopicRow).toBeInTheDocument(); const extBtn = within(externalTopicRow).getByRole('button', { diff --git a/kafka-ui-react-app/src/components/common/BytesFormatted/BytesFormatted.styled.ts b/kafka-ui-react-app/src/components/common/BytesFormatted/BytesFormatted.styled.ts new file mode 100644 index 00000000000..dc24760572b --- /dev/null +++ b/kafka-ui-react-app/src/components/common/BytesFormatted/BytesFormatted.styled.ts @@ -0,0 +1,5 @@ +import styled from 'styled-components'; + +export const NoWrap = styled.span` + white-space: nowrap; +`; diff --git a/kafka-ui-react-app/src/components/common/BytesFormatted/BytesFormatted.tsx b/kafka-ui-react-app/src/components/common/BytesFormatted/BytesFormatted.tsx index 7cfb35cfd21..5bfc5a1670e 100644 --- a/kafka-ui-react-app/src/components/common/BytesFormatted/BytesFormatted.tsx +++ b/kafka-ui-react-app/src/components/common/BytesFormatted/BytesFormatted.tsx @@ -1,5 +1,7 @@ import React from 'react'; +import { NoWrap } from './BytesFormatted.styled'; + interface Props { value: string | number | undefined; precision?: number; @@ -8,22 +10,22 @@ interface Props { export const sizes = ['Bytes', 'KB', 'MB', 'GB', 'TB', 'PB', 'EB', 'ZB', 'YB']; const BytesFormatted: React.FC<Props> = ({ value, precision = 0 }) => { - const formatedValue = React.useMemo((): string => { + const formattedValue = React.useMemo((): string => { try { const bytes = typeof value === 'string' ? parseInt(value, 10) : value; if (Number.isNaN(bytes) || (bytes && bytes < 0)) return `-Bytes`; - if (!bytes || bytes < 1024) return `${Math.ceil(bytes || 0)}${sizes[0]}`; + if (!bytes || bytes < 1024) return `${Math.ceil(bytes || 0)} ${sizes[0]}`; const pow = Math.floor(Math.log2(bytes) / 10); const multiplier = 10 ** (precision < 0 ? 0 : precision); - return ( - Math.round((bytes * multiplier) / 1024 ** pow) / multiplier + sizes[pow] - ); + return `${Math.round((bytes * multiplier) / 1024 ** pow) / multiplier} ${ + sizes[pow] + }`; } catch (e) { return `-Bytes`; } }, [precision, value]); - return <span>{formatedValue}</span>; + return <NoWrap>{formattedValue}</NoWrap>; }; export default BytesFormatted; diff --git a/kafka-ui-react-app/src/components/common/BytesFormatted/__tests__/BytesFormatted.spec.tsx b/kafka-ui-react-app/src/components/common/BytesFormatted/__tests__/BytesFormatted.spec.tsx index bfb7b85f502..e97767fa201 100644 --- a/kafka-ui-react-app/src/components/common/BytesFormatted/__tests__/BytesFormatted.spec.tsx +++ b/kafka-ui-react-app/src/components/common/BytesFormatted/__tests__/BytesFormatted.spec.tsx @@ -7,29 +7,29 @@ import { render, screen } from '@testing-library/react'; describe('BytesFormatted', () => { it('renders Bytes correctly', () => { render(<BytesFormatted value={666} />); - expect(screen.getByText('666Bytes')).toBeInTheDocument(); + expect(screen.getByText('666 Bytes')).toBeInTheDocument(); }); it('renders correct units', () => { let value = 1; sizes.forEach((unit) => { render(<BytesFormatted value={value} />); - expect(screen.getByText(`1${unit}`)).toBeInTheDocument(); + expect(screen.getByText(`1 ${unit}`)).toBeInTheDocument(); value *= 1024; }); }); it('renders correct precision', () => { render(<BytesFormatted value={2000} precision={100} />); - expect(screen.getByText(`1.953125${sizes[1]}`)).toBeInTheDocument(); + expect(screen.getByText(`1.953125 ${sizes[1]}`)).toBeInTheDocument(); render(<BytesFormatted value={10000} precision={5} />); - expect(screen.getByText(`9.76563${sizes[1]}`)).toBeInTheDocument(); + expect(screen.getByText(`9.76563 ${sizes[1]}`)).toBeInTheDocument(); }); it('correctly handles invalid props', () => { render(<BytesFormatted value={10000} precision={-1} />); - expect(screen.getByText(`10${sizes[1]}`)).toBeInTheDocument(); + expect(screen.getByText(`10 ${sizes[1]}`)).toBeInTheDocument(); render(<BytesFormatted value="some string" />); expect(screen.getAllByText(`-${sizes[0]}`).length).toBeTruthy(); @@ -38,6 +38,6 @@ describe('BytesFormatted', () => { expect(screen.getAllByText(`-${sizes[0]}`).length).toBeTruthy(); render(<BytesFormatted value={undefined} />); - expect(screen.getByText(`0${sizes[0]}`)).toBeInTheDocument(); + expect(screen.getByText(`0 ${sizes[0]}`)).toBeInTheDocument(); }); }); diff --git a/kafka-ui-react-app/src/components/common/NewTable/__test__/Table.spec.tsx b/kafka-ui-react-app/src/components/common/NewTable/__test__/Table.spec.tsx index dbfdddab662..2c65ea74b3d 100644 --- a/kafka-ui-react-app/src/components/common/NewTable/__test__/Table.spec.tsx +++ b/kafka-ui-react-app/src/components/common/NewTable/__test__/Table.spec.tsx @@ -125,10 +125,10 @@ describe('Table', () => { it('renders SizeCell', () => { renderComponent(); - expect(screen.getByText('1KB')).toBeInTheDocument(); - expect(screen.getByText('3Bytes')).toBeInTheDocument(); - expect(screen.getByText('49KB')).toBeInTheDocument(); - expect(screen.getByText('1MB')).toBeInTheDocument(); + expect(screen.getByText('1 KB')).toBeInTheDocument(); + expect(screen.getByText('3 Bytes')).toBeInTheDocument(); + expect(screen.getByText('49 KB')).toBeInTheDocument(); + expect(screen.getByText('1 MB')).toBeInTheDocument(); }); it('renders TimestampCell', () => {
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java index bf618f0222a..d707b5acafd 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java @@ -175,7 +175,7 @@ void produceMessage() { .openTopic(TOPIC_FOR_MESSAGES.getName()); topicDetails .waitUntilScreenReady() - .openTopicMenu(TopicDetails.TopicMenu.MESSAGES) + .openDetailsTab(TopicDetails.TopicMenu.MESSAGES) .clickProduceMessageBtn(); produceMessagePanel .waitUntilScreenReady() @@ -204,7 +204,7 @@ void clearMessage() { .openTopic(TOPIC_FOR_MESSAGES.getName()); topicDetails .waitUntilScreenReady() - .openTopicMenu(TopicDetails.TopicMenu.OVERVIEW) + .openDetailsTab(TopicDetails.TopicMenu.OVERVIEW) .clickProduceMessageBtn(); produceMessagePanel .waitUntilScreenReady() @@ -238,7 +238,7 @@ void redirectToConsumerFromTopic() { .openTopic(topicName); topicDetails .waitUntilScreenReady() - .openTopicMenu(TopicDetails.TopicMenu.CONSUMERS) + .openDetailsTab(TopicDetails.TopicMenu.CONSUMERS) .openConsumerGroup(consumerGroupId); consumersDetails .waitUntilScreenReady();
train
val
2022-11-07T12:41:04
"2022-09-30T09:22:06Z"
Haarolean
train
provectus/kafka-ui/2822_2849
provectus/kafka-ui
provectus/kafka-ui/2822
provectus/kafka-ui/2849
[ "connected" ]
77f789fa6c4a2471417cf0136788ad39952df808
cd2d8b026ffb8ec4ba9314dffb63465e50fe59b6
[]
[]
"2022-10-30T19:54:49Z"
[ "good first issue", "scope/frontend", "status/accepted", "type/chore", "hacktoberfest" ]
KSQL: Container with query "Status" is too narrow
**Set up** [d1d46b6](https://github.com/provectus/kafka-ui/commit/d1d46b6) **Steps to Reproduce** 1. Open KSQL DB in navigation sidebar. 2. Fill in some query and Execute it. E.g.: `CREATE STREAM first.messagesStream1 (viewtime BIGINT, userid VARCHAR, pageid VARCHAR) WITH (KAFKA_TOPIC='first.messages', VALUE_FORMAT='DELIMITED');` 4. See the status bar under the "Execute", "Stop query", and "Clear results" button. **Actual behavior** Status container is too narrow, when the window width exceeds 1200px: https://user-images.githubusercontent.com/112083452/197765749-8671d584-cf43-4fc4-9680-6c88e50b560e.mov **Expected behavior** All containers should fit window width without any overlay.
[ "kafka-ui-react-app/src/components/KsqlDb/Query/renderer/TableRenderer/TableRenderer.styled.tsx" ]
[ "kafka-ui-react-app/src/components/KsqlDb/Query/renderer/TableRenderer/TableRenderer.styled.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/KsqlDb/Query/renderer/TableRenderer/TableRenderer.styled.tsx b/kafka-ui-react-app/src/components/KsqlDb/Query/renderer/TableRenderer/TableRenderer.styled.tsx index c02c30d3691..96ef549aa69 100644 --- a/kafka-ui-react-app/src/components/KsqlDb/Query/renderer/TableRenderer/TableRenderer.styled.tsx +++ b/kafka-ui-react-app/src/components/KsqlDb/Query/renderer/TableRenderer/TableRenderer.styled.tsx @@ -8,6 +8,7 @@ export const Wrapper = styled.div` export const ScrollableTable = styled(Table)` overflow-y: scroll; + width: 100%; td { vertical-align: top;
null
train
val
2022-11-01T10:27:04
"2022-10-25T11:50:43Z"
BulatKha
train
provectus/kafka-ui/2712_2851
provectus/kafka-ui
provectus/kafka-ui/2712
provectus/kafka-ui/2851
[ "connected" ]
cd2d8b026ffb8ec4ba9314dffb63465e50fe59b6
ee920b43c45156cad1b89be31e43106ca3859edb
[]
[ "don't we want gitBtn and discordBtn to be visible?", "not sure that we really need to search discordBtn with '4DWzD7pGE5', test will fail everytime it'll be changed. let's create locator using 'contains' w/o discord's id", "added to getAllVisibleElements ", "Refactored " ]
"2022-10-31T14:28:54Z"
[ "scope/QA", "scope/AQA" ]
[e2e] refactoring of SmokeTest
- current test about comparing screenshots is redundant - need to create test which will check all base elements, such as logo, naviPanel availability, logout button
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/NaviSideBar.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/screenshots/NoReferenceScreenshotFoundException.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/screenshots/Screenshooter.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/NaviSideBar.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/TopPanel.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/SmokeTests.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/BaseTest.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/NaviSideBar.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/NaviSideBar.java index 4e6bab8eb00..0f6f4c22609 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/NaviSideBar.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/NaviSideBar.java @@ -5,6 +5,9 @@ import io.qameta.allure.Step; import java.time.Duration; +import java.util.List; +import java.util.stream.Collectors; +import java.util.stream.Stream; import static com.codeborne.selenide.Selenide.$x; import static com.provectus.kafka.ui.settings.Source.CLUSTER_NAME; @@ -16,6 +19,14 @@ public class NaviSideBar { protected String sideMenuOptionElementLocator = ".//ul/li[contains(.,'%s')]"; protected String clusterElementLocator = "//aside/ul/li[contains(.,'%s')]"; + private SelenideElement expandCluster(String clusterName) { + SelenideElement clusterElement = $x(String.format(clusterElementLocator, clusterName)).shouldBe(Condition.visible); + if (clusterElement.parent().$$x(".//ul").size() == 0) { + clusterElement.click(); + } + return clusterElement; + } + @Step public NaviSideBar waitUntilScreenReady() { loadingSpinner.shouldBe(Condition.disappear, Duration.ofSeconds(30)); @@ -25,11 +36,7 @@ public NaviSideBar waitUntilScreenReady() { @Step public NaviSideBar openSideMenu(String clusterName, SideMenuOption option) { - SelenideElement clusterElement = $x(String.format(clusterElementLocator, clusterName)).shouldBe(Condition.visible); - if (clusterElement.parent().$$x(".//ul").size() == 0) { - clusterElement.click(); - } - clusterElement + expandCluster(clusterName) .parent() .$x(String.format(sideMenuOptionElementLocator, option.value)) .click(); @@ -43,6 +50,7 @@ public NaviSideBar openSideMenu(SideMenuOption option) { } public enum SideMenuOption { + DASHBOARD("Dashboard"), BROKERS("Brokers"), TOPICS("Topics"), CONSUMERS("Consumers"), @@ -56,4 +64,11 @@ public enum SideMenuOption { this.value = value; } } + + public List<SelenideElement> getAllMenuButtons() { + expandCluster(CLUSTER_NAME); + return Stream.of(SideMenuOption.values()) + .map(option -> $x(String.format(sideMenuOptionElementLocator, option.value))) + .collect(Collectors.toList()); + } } \ No newline at end of file diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/TopPanel.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/TopPanel.java new file mode 100644 index 00000000000..8101543abfc --- /dev/null +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/TopPanel.java @@ -0,0 +1,24 @@ +package com.provectus.kafka.ui.pages; + +import com.codeborne.selenide.SelenideElement; + +import java.util.Arrays; +import java.util.List; + +import static com.codeborne.selenide.Selenide.$x; + +public class TopPanel { + protected SelenideElement kafkaLogo = $x("//a[contains(text(),'UI for Apache Kafka')]"); + protected SelenideElement kafkaVersion = $x("//a[@title='Current commit']"); + protected SelenideElement logOutBtn = $x("//button[contains(text(),'Log out')]"); + protected SelenideElement gitBtn = $x("//a[@href='https://github.com/provectus/kafka-ui']"); + protected SelenideElement discordBtn = $x("//a[contains(@href,'https://discord.com/invite')]"); + + public List<SelenideElement> getAllVisibleElements() { + return Arrays.asList(kafkaLogo, kafkaVersion, logOutBtn, gitBtn, discordBtn); + } + + public List<SelenideElement> getAllEnabledElements() { + return Arrays.asList(logOutBtn, gitBtn, discordBtn, kafkaLogo); + } +} diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/screenshots/NoReferenceScreenshotFoundException.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/screenshots/NoReferenceScreenshotFoundException.java deleted file mode 100644 index 0011c4f0a13..00000000000 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/screenshots/NoReferenceScreenshotFoundException.java +++ /dev/null @@ -1,7 +0,0 @@ -package com.provectus.kafka.ui.utilities.screenshots; - -public class NoReferenceScreenshotFoundException extends Throwable { - public NoReferenceScreenshotFoundException(String name) { - super("no reference screenshot found for " + name); - } -} diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/screenshots/Screenshooter.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/screenshots/Screenshooter.java deleted file mode 100644 index 136a9e8bd4b..00000000000 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/screenshots/Screenshooter.java +++ /dev/null @@ -1,162 +0,0 @@ -package com.provectus.kafka.ui.utilities.screenshots; - -import io.qameta.allure.Allure; -import io.qameta.allure.Attachment; -import lombok.SneakyThrows; -import org.junit.jupiter.api.Assertions; -import org.slf4j.Logger; -import org.slf4j.LoggerFactory; -import ru.yandex.qatools.ashot.AShot; -import ru.yandex.qatools.ashot.Screenshot; -import ru.yandex.qatools.ashot.comparison.ImageDiff; -import ru.yandex.qatools.ashot.comparison.ImageDiffer; -import ru.yandex.qatools.ashot.coordinates.WebDriverCoordsProvider; - -import javax.imageio.ImageIO; -import java.awt.image.BufferedImage; -import java.io.ByteArrayOutputStream; -import java.io.File; -import java.io.IOException; -import java.nio.file.FileSystems; -import java.util.List; - -import static com.codeborne.selenide.WebDriverRunner.getWebDriver; - -public class Screenshooter { - - public static Logger log = LoggerFactory.getLogger(Screenshooter.class); - - private static final int PIXELS_THRESHOLD = - Integer.parseInt(System.getProperty("PIXELS_THRESHOLD", "200")); - private static final String SCREENSHOTS_FOLDER = - System.getProperty("SCREENSHOTS_FOLDER", "com/provectus/kafka/ui/utilities/screenshots/"); - private static final String DIFF_SCREENSHOTS_FOLDER = - System.getProperty("DIFF_SCREENSHOTS_FOLDER", "build/__diff__/"); - private static final String ACTUAL_SCREENSHOTS_FOLDER = - System.getProperty("ACTUAL_SCREENSHOTS_FOLDER", "build/__actual__/"); - private static final boolean SHOULD_SAVE_SCREENSHOTS_IF_NOT_EXIST = - Boolean.parseBoolean(System.getProperty("SHOULD_SAVE_SCREENSHOTS_IF_NOT_EXIST", "true")); - private static final boolean TURN_OFF_SCREENSHOTS = - Boolean.parseBoolean(System.getProperty("TURN_OFF_SCREENSHOTS", "false")); - private static final boolean USE_LOCAL_BROWSER = - Boolean.parseBoolean(System.getProperty("USE_LOCAL_BROWSER", "false")); - - private File newFile(String name) { - var file = new File(name); - if (!file.exists()) { - file.mkdirs(); - } - return file; - } - - public Screenshooter() { - List.of(SCREENSHOTS_FOLDER, DIFF_SCREENSHOTS_FOLDER, ACTUAL_SCREENSHOTS_FOLDER) - .forEach(this::newFile); - } - - public void compareScreenshots(String name) { - compareScreenshots(name, false); - } - - - public void compareScreenshots(String name, boolean shouldUpdateScreenshotIfDiffer) { - if (TURN_OFF_SCREENSHOTS || USE_LOCAL_BROWSER) { - log.warn(String.format("compareScreenshots turned off due TURN_OFF_SCREENSHOTS || USE_LOCAL_BROWSER: %b || %b" - , TURN_OFF_SCREENSHOTS, USE_LOCAL_BROWSER)); - return; - } - if (!doesScreenshotExist(name)) { - if (SHOULD_SAVE_SCREENSHOTS_IF_NOT_EXIST) { - updateActualScreenshot(name); - } else { - try { - throw new NoReferenceScreenshotFoundException(name); - } catch (NoReferenceScreenshotFoundException e) { - e.printStackTrace(); - } - } - } else { - makeImageDiff(name, shouldUpdateScreenshotIfDiffer); - } - } - - - private void updateActualScreenshot(String name) { - Screenshot actual = - new AShot().coordsProvider(new WebDriverCoordsProvider()).takeScreenshot(getWebDriver()); - File file = newFile(SCREENSHOTS_FOLDER + name + ".png"); - try { - ImageIO.write(actual.getImage(), "png", file); - } catch (IOException e) { - e.printStackTrace(); - } - log.debug(String.format("created screenshot: %s \n at %s", name, file.getAbsolutePath())); - } - - private static boolean doesScreenshotExist(String name) { - return new File(SCREENSHOTS_FOLDER + name + ".png").exists(); - } - - @SneakyThrows - private void makeImageDiff(String expectedName, boolean shouldUpdateScreenshotIfDiffer) { - String fullPathNameExpected = SCREENSHOTS_FOLDER + expectedName + ".png"; - String fullPathNameActual = ACTUAL_SCREENSHOTS_FOLDER + expectedName + ".png"; - String fullPathNameDiff = DIFF_SCREENSHOTS_FOLDER + expectedName + ".png"; - - // activating allure plugin for showing diffs in report - Allure.label("testType", "screenshotDiff"); - - Screenshot actual = - new AShot().coordsProvider(new WebDriverCoordsProvider()).takeScreenshot(getWebDriver()); - ImageIO.write(actual.getImage(), "png", newFile(fullPathNameActual)); - - Screenshot expected = new Screenshot(ImageIO.read(newFile(fullPathNameExpected))); - ImageDiff diff = new ImageDiffer().makeDiff(actual, expected); - BufferedImage diffImage = diff.getMarkedImage(); - ImageIO.write(diffImage, "png", newFile(fullPathNameDiff)); - // adding to report - diff(fullPathNameDiff); - // adding to report - actual(fullPathNameActual); - // adding to report - expected(fullPathNameExpected); - - if (shouldUpdateScreenshotIfDiffer) { - if (diff.getDiffSize() > PIXELS_THRESHOLD) { - updateActualScreenshot(expectedName); - } - } else { - Assertions.assertTrue( - PIXELS_THRESHOLD >= diff.getDiffSize(), - String.format("Amount of differing pixels should be less or equals than %s, actual %s\n" + - "diff file: %s", - PIXELS_THRESHOLD, diff.getDiffSize(), FileSystems.getDefault().getPath(fullPathNameDiff).normalize().toAbsolutePath())); - } - } - - @SneakyThrows - private byte[] imgToBytes(String filename) { - BufferedImage bImage2 = ImageIO.read(new File(filename)); - var bos2 = new ByteArrayOutputStream(); - ImageIO.write(bImage2, "png", bos2); - return bos2.toByteArray(); - } - - - @Attachment - private byte[] actual(String actualFileName) { - return imgToBytes(actualFileName); - } - - - @Attachment - private byte[] expected(String expectedFileName) { - return imgToBytes(expectedFileName); - } - - - @Attachment - private byte[] diff(String diffFileName) { - return imgToBytes(diffFileName); - } -}
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/SmokeTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/SmokeTests.java index e876d3a634b..e1e6927c432 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/SmokeTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/SmokeTests.java @@ -1,10 +1,11 @@ package com.provectus.kafka.ui; +import com.codeborne.selenide.Condition; import com.provectus.kafka.ui.base.BaseTest; import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.AutomationStatus; import com.provectus.kafka.ui.utilities.qaseIoUtils.enums.Status; import io.qase.api.annotation.CaseId; -import org.junit.jupiter.api.DisplayName; +import org.assertj.core.api.SoftAssertions; import org.junit.jupiter.api.Test; public class SmokeTests extends BaseTest { @@ -12,8 +13,20 @@ public class SmokeTests extends BaseTest { @Test @AutomationStatus(status = Status.AUTOMATED) @CaseId(198) - @DisplayName("main page should load") - void mainPageLoads() { - compareScreenshots("main"); + public void checkBasePageElements(){ + SoftAssertions softly = new SoftAssertions(); + topPanel.getAllVisibleElements() + .forEach(element -> + softly.assertThat(element.is(Condition.visible)) + .as(element.getSearchCriteria() + " isVisible()").isTrue()); + topPanel.getAllEnabledElements() + .forEach(element -> + softly.assertThat(element.is(Condition.enabled)) + .as(element.getSearchCriteria() + " isEnabled()").isTrue()); + naviSideBar.getAllMenuButtons() + .forEach(element -> + softly.assertThat(element.is(Condition.enabled) && element.is(Condition.visible)) + .as(element.getSearchCriteria() + " isEnabled()").isTrue()); + softly.assertAll(); } -} +} \ No newline at end of file diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/BaseTest.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/BaseTest.java index a3086fd7151..11cf07c622e 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/BaseTest.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/BaseTest.java @@ -1,11 +1,9 @@ package com.provectus.kafka.ui.base; -import com.codeborne.selenide.Condition; import com.codeborne.selenide.Selenide; import com.codeborne.selenide.WebDriverRunner; import com.provectus.kafka.ui.utilities.qaseIoUtils.DisplayNameGenerator; import com.provectus.kafka.ui.utilities.qaseIoUtils.TestCaseGenerator; -import com.provectus.kafka.ui.utilities.screenshots.Screenshooter; import io.github.cdimascio.dotenv.Dotenv; import io.qameta.allure.Allure; import lombok.extern.slf4j.Slf4j; @@ -24,7 +22,6 @@ import java.io.ByteArrayInputStream; import java.io.File; import java.io.IOException; -import java.time.Duration; import static com.provectus.kafka.ui.base.Setup.*; import static com.provectus.kafka.ui.settings.Source.BASE_WEB_URL; @@ -35,18 +32,8 @@ public class BaseTest extends Facade { private static final String SELENIUM_IMAGE_NAME = "selenium/standalone-chrome:103.0"; private static final String SELENIARM_STANDALONE_CHROMIUM = "seleniarm/standalone-chromium:103.0"; - private final Screenshooter screenshooter = new Screenshooter(); - protected static BrowserWebDriverContainer<?> webDriverContainer = null; - public void compareScreenshots(String name) { - screenshooter.compareScreenshots(name); - } - - public void compareScreenshots(String name, Boolean shouldUpdateScreenshots) { - screenshooter.compareScreenshots(name, shouldUpdateScreenshots); - } - private static boolean isARM64() { return System.getProperty("os.arch").equals("aarch64"); } diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java index ac223c62b11..5b9533c30ec 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java @@ -2,6 +2,7 @@ import com.provectus.kafka.ui.helpers.ApiHelper; import com.provectus.kafka.ui.pages.NaviSideBar; +import com.provectus.kafka.ui.pages.TopPanel; import com.provectus.kafka.ui.pages.connector.ConnectorCreateForm; import com.provectus.kafka.ui.pages.connector.ConnectorDetails; import com.provectus.kafka.ui.pages.connector.KafkaConnectList; @@ -30,4 +31,5 @@ public abstract class Facade { protected ConsumersDetails consumersDetails = new ConsumersDetails(); protected ConsumersList consumersList = new ConsumersList(); protected NaviSideBar naviSideBar = new NaviSideBar(); + protected TopPanel topPanel = new TopPanel(); }
train
val
2022-11-01T13:20:43
"2022-10-07T12:01:22Z"
VladSenyuta
train
provectus/kafka-ui/2672_2854
provectus/kafka-ui
provectus/kafka-ui/2672
provectus/kafka-ui/2854
[ "connected" ]
734d4ccdf71ee9b5c85570faa0c4767db3624509
5efb380c42a97418b29fd5e72ed8372be7d1b48e
[ "Hello @gabrieljones, thank you for suggestion.\r\n\r\nI personally find this behaviour useful, especially when you need to send similar messages with minor changes. Maybe we can add explicit \"Generate random\" button in produce dialog. \r\nWe will discuss it internally and implement it either way. \r\n\r\ncc @Haarolean @workshur \r\n", "To do:\r\n1. In the produce message pane add a new checkbox \"Keep the contents\", unchecked by default\r\n2. If the checkbox is being checked, upon producing a message do the following:\r\n1) do not close the pane upon producing the message\r\n2) Do not alter key/value content in any way" ]
[ "do we need the `div` tag here ?", "isn't submit happening any how , `if (!keepContents) {}` should only act on the values without the `onSubmit` where it should be outside of the condition, cause the request is already being carried.", "@Mgrdich this function is not mine, as I understand it, it only closes the sidebar, I will rename it", "we do not need div", "then we can delete it.", "yeah i want to understand when we should trigger that close thing , cause from only reading of it , the request the BE is happening -> so logically `onSubmit` should work, `if (!keepContents) {` `setValues` is what is happening to the ui state of the component.\r\n\r\nif it is not onSubmit handler then it should be renamed" ]
"2022-11-01T06:24:16Z"
[ "type/enhancement", "good first issue", "scope/frontend", "status/accepted" ]
Don't Dismiss Produce Message pane after `Send`
### Which version of the app are you running? vv0.4.0([521ba0c](https://github.com/provectus/kafka-ui/commit/521ba0c)) ### Is your proposal related to a problem? After clicking `Send` in Produce Message pane UI shift back to the Messages pane and resets the Produce Message field contents. ### Describe the solution you'd like After clicking `Send` in Produce Message pane UI stays on the Produce Message pane and the field contents remain as entered without being reset to a new random lorem ipsum.
[ "kafka-ui-react-app/src/components/Topics/Topic/SendMessage/SendMessage.styled.tsx", "kafka-ui-react-app/src/components/Topics/Topic/SendMessage/SendMessage.tsx", "kafka-ui-react-app/src/components/Topics/Topic/SendMessage/__test__/SendMessage.spec.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Topic.tsx", "kafka-ui-react-app/src/components/common/SlidingSidebar/SlidingSidebar.styled.ts" ]
[ "kafka-ui-react-app/src/components/Topics/Topic/SendMessage/SendMessage.styled.tsx", "kafka-ui-react-app/src/components/Topics/Topic/SendMessage/SendMessage.tsx", "kafka-ui-react-app/src/components/Topics/Topic/SendMessage/__test__/SendMessage.spec.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Topic.tsx", "kafka-ui-react-app/src/components/common/SlidingSidebar/SlidingSidebar.styled.ts" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/Topic/SendMessage/SendMessage.styled.tsx b/kafka-ui-react-app/src/components/Topics/Topic/SendMessage/SendMessage.styled.tsx index 483c41d0533..d2750abf7d1 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/SendMessage/SendMessage.styled.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/SendMessage/SendMessage.styled.tsx @@ -8,15 +8,29 @@ export const Wrapper = styled.div` export const Columns = styled.div` margin: -0.75rem; margin-bottom: 0.75rem; + display: flex; + flex-direction: column; + padding: 0.75rem; + gap: 8px; @media screen and (min-width: 769px) { display: flex; } `; - -export const Column = styled.div` - flex-basis: 0; - flex-grow: 1; - flex-shrink: 1; - padding: 0.75rem; +export const Flex = styled.div` + display: flex; + flex-direction: row; + gap: 8px; + @media screen and (max-width: 1200px) { + flex-direction: column; + } +`; +export const FlexItem = styled.div` + width: 18rem; + @media screen and (max-width: 1450px) { + width: 50%; + } + @media screen and (max-width: 1200px) { + width: 100%; + } `; diff --git a/kafka-ui-react-app/src/components/Topics/Topic/SendMessage/SendMessage.tsx b/kafka-ui-react-app/src/components/Topics/Topic/SendMessage/SendMessage.tsx index 9450e512ad1..bacfa76c93f 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/SendMessage/SendMessage.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/SendMessage/SendMessage.tsx @@ -4,6 +4,7 @@ import { RouteParamsClusterTopic } from 'lib/paths'; import { Button } from 'components/common/Button/Button'; import Editor from 'components/common/Editor/Editor'; import Select, { SelectOption } from 'components/common/Select/Select'; +import Switch from 'components/common/Switch/Switch'; import useAppParams from 'lib/hooks/useAppParams'; import { showAlert } from 'lib/errorHandling'; import { useSendMessage, useTopicDetails } from 'lib/hooks/api/topics'; @@ -26,9 +27,12 @@ interface FormType { partition: number; keySerde: string; valueSerde: string; + keepContents: boolean; } -const SendMessage: React.FC<{ onSubmit: () => void }> = ({ onSubmit }) => { +const SendMessage: React.FC<{ closeSidebar: () => void }> = ({ + closeSidebar, +}) => { const { clusterName, topicName } = useAppParams<RouteParamsClusterTopic>(); const { data: topic } = useTopicDetails({ clusterName, topicName }); const { data: serdes = {} } = useSerdes({ @@ -47,11 +51,13 @@ const SendMessage: React.FC<{ onSubmit: () => void }> = ({ onSubmit }) => { handleSubmit, formState: { isSubmitting }, control, + setValue, } = useForm<FormType>({ mode: 'onChange', defaultValues: { ...defaultValues, partition: Number(partitionOptions[0].value), + keepContents: false, }, }); @@ -62,6 +68,7 @@ const SendMessage: React.FC<{ onSubmit: () => void }> = ({ onSubmit }) => { content, headers, partition, + keepContents, }: FormType) => { let errors: string[] = []; @@ -110,7 +117,11 @@ const SendMessage: React.FC<{ onSubmit: () => void }> = ({ onSubmit }) => { keySerde, valueSerde, }); - onSubmit(); + if (!keepContents) { + setValue('key', ''); + setValue('content', ''); + closeSidebar(); + } } catch (e) { // do nothing } @@ -120,7 +131,7 @@ const SendMessage: React.FC<{ onSubmit: () => void }> = ({ onSubmit }) => { <S.Wrapper> <form onSubmit={handleSubmit(submit)}> <S.Columns> - <S.Column> + <S.FlexItem> <InputLabel>Partition</InputLabel> <Controller control={control} @@ -137,47 +148,58 @@ const SendMessage: React.FC<{ onSubmit: () => void }> = ({ onSubmit }) => { /> )} /> - </S.Column> - <S.Column> - <InputLabel>Key Serde</InputLabel> + </S.FlexItem> + <S.Flex> + <S.FlexItem> + <InputLabel>Key Serde</InputLabel> + <Controller + control={control} + name="keySerde" + render={({ field: { name, onChange, value } }) => ( + <Select + id="selectKeySerdeOptions" + aria-labelledby="selectKeySerdeOptions" + name={name} + onChange={onChange} + minWidth="100%" + options={getSerdeOptions(serdes.key || [])} + value={value} + /> + )} + /> + </S.FlexItem> + <S.FlexItem> + <InputLabel>Value Serde</InputLabel> + <Controller + control={control} + name="valueSerde" + render={({ field: { name, onChange, value } }) => ( + <Select + id="selectValueSerdeOptions" + aria-labelledby="selectValueSerdeOptions" + name={name} + onChange={onChange} + minWidth="100%" + options={getSerdeOptions(serdes.value || [])} + value={value} + /> + )} + /> + </S.FlexItem> + </S.Flex> + <div> <Controller control={control} - name="keySerde" + name="keepContents" render={({ field: { name, onChange, value } }) => ( - <Select - id="selectKeySerdeOptions" - aria-labelledby="selectKeySerdeOptions" - name={name} - onChange={onChange} - minWidth="100%" - options={getSerdeOptions(serdes.key || [])} - value={value} - /> + <Switch name={name} onChange={onChange} checked={value} /> )} /> - </S.Column> - <S.Column> - <InputLabel>Value Serde</InputLabel> - <Controller - control={control} - name="valueSerde" - render={({ field: { name, onChange, value } }) => ( - <Select - id="selectValueSerdeOptions" - aria-labelledby="selectValueSerdeOptions" - name={name} - onChange={onChange} - minWidth="100%" - options={getSerdeOptions(serdes.value || [])} - value={value} - /> - )} - /> - </S.Column> + <InputLabel>Keep contents</InputLabel> + </div> </S.Columns> - <S.Columns> - <S.Column> + <div> <InputLabel>Key</InputLabel> <Controller control={control} @@ -191,8 +213,8 @@ const SendMessage: React.FC<{ onSubmit: () => void }> = ({ onSubmit }) => { /> )} /> - </S.Column> - <S.Column> + </div> + <div> <InputLabel>Value</InputLabel> <Controller control={control} @@ -206,10 +228,10 @@ const SendMessage: React.FC<{ onSubmit: () => void }> = ({ onSubmit }) => { /> )} /> - </S.Column> + </div> </S.Columns> <S.Columns> - <S.Column> + <div> <InputLabel>Headers</InputLabel> <Controller control={control} @@ -224,7 +246,7 @@ const SendMessage: React.FC<{ onSubmit: () => void }> = ({ onSubmit }) => { /> )} /> - </S.Column> + </div> </S.Columns> <Button buttonSize="M" diff --git a/kafka-ui-react-app/src/components/Topics/Topic/SendMessage/__test__/SendMessage.spec.tsx b/kafka-ui-react-app/src/components/Topics/Topic/SendMessage/__test__/SendMessage.spec.tsx index 609324b8f09..35137ef8643 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/SendMessage/__test__/SendMessage.spec.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/SendMessage/__test__/SendMessage.spec.tsx @@ -49,7 +49,7 @@ const renderComponent = async () => { const path = clusterTopicPath(clusterName, topicName); await render( <WithRoute path={clusterTopicPath()}> - <SendMessage onSubmit={mockOnSubmit} /> + <SendMessage closeSidebar={mockOnSubmit} /> </WithRoute>, { initialEntries: [path] } ); diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Topic.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Topic.tsx index 8945523576b..0022e002a73 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Topic.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Topic.tsx @@ -236,7 +236,7 @@ const Topic: React.FC = () => { title="Produce Message" > <Suspense fallback={<PageLoader />}> - <SendMessage onSubmit={closeSidebar} /> + <SendMessage closeSidebar={closeSidebar} /> </Suspense> </SlidingSidebar> </> diff --git a/kafka-ui-react-app/src/components/common/SlidingSidebar/SlidingSidebar.styled.ts b/kafka-ui-react-app/src/components/common/SlidingSidebar/SlidingSidebar.styled.ts index 710a85a86c1..4ccb4ae0b73 100644 --- a/kafka-ui-react-app/src/components/common/SlidingSidebar/SlidingSidebar.styled.ts +++ b/kafka-ui-react-app/src/components/common/SlidingSidebar/SlidingSidebar.styled.ts @@ -6,7 +6,7 @@ export const Wrapper = styled.div<{ $open?: boolean }>( position: fixed; top: ${theme.layout.navBarHeight}; bottom: 0; - width: 60vw; + width: 37vw; right: calc(${$open ? '0px' : theme.layout.rightSidebarWidth} * -1); box-shadow: -1px 0px 10px 0px rgba(0, 0, 0, 0.2); transition: right 0.3s linear;
null
train
val
2023-04-20T19:13:05
"2022-09-30T03:35:47Z"
gabrieljones
train
provectus/kafka-ui/2488_2862
provectus/kafka-ui
provectus/kafka-ui/2488
provectus/kafka-ui/2862
[ "connected" ]
5900f8e957f2abf078bb6db4552b186a9e21a0b5
33784e3e837a9f64dc4a39f4147e1c11e54f9401
[]
[]
"2022-11-01T20:20:15Z"
[ "type/bug", "scope/frontend", "status/accepted", "status/confirmed", "type/regression" ]
Messages pagination is missing
backward button -- use as browser's "back" function forward -- as it was before use /v2 as reference
[ "kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/Messages.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/__test__/MessagesTable.spec.tsx" ]
[ "kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/Messages.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/__test__/MessagesTable.spec.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx index c010ca65cb5..87d59f3b8e3 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx @@ -93,6 +93,8 @@ const Filters: React.FC<FiltersProps> = ({ const navigate = useNavigate(); const [searchParams] = useSearchParams(); + const page = searchParams.get('page'); + const { data: topic } = useTopicDetails({ clusterName, topicName }); const partitions = topic?.partitions || []; @@ -198,6 +200,7 @@ const Filters: React.FC<FiltersProps> = ({ filterQueryType: queryType, attempt: nextAttempt, limit: PER_PAGE, + page: page || 0, seekDirection, keySerde: keySerde || (searchParams.get('keySerde') as string), valueSerde: valueSerde || (searchParams.get('valueSerde') as string), @@ -379,6 +382,7 @@ const Filters: React.FC<FiltersProps> = ({ timestamp, query, seekDirection, + page, ]); React.useEffect(() => { diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Messages.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Messages.tsx index b71e4eb8fa2..35490c627e3 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Messages.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Messages.tsx @@ -6,6 +6,7 @@ import { useSerdes } from 'lib/hooks/api/topicMessages'; import useAppParams from 'lib/hooks/useAppParams'; import { RouteParamsClusterTopic } from 'lib/paths'; import { getDefaultSerdeName } from 'components/Topics/Topic/Messages/getDefaultSerdeName'; +import { MESSAGES_PER_PAGE } from 'lib/constants'; import MessagesTable from './MessagesTable'; import FiltersContainer from './Filters/FiltersContainer'; @@ -47,6 +48,9 @@ const Messages: React.FC = () => { if (!searchParams.get('valueSerde')) { searchParams.set('valueSerde', getDefaultSerdeName(serdes.value || [])); } + if (!searchParams.get('limit')) { + searchParams.set('limit', MESSAGES_PER_PAGE); + } setSearchParams(searchParams); }, [serdes]); diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx index 3c5901214b0..96e9f40d73d 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx @@ -9,6 +9,10 @@ import { } from 'redux/reducers/topicMessages/selectors'; import TopicMessagesContext from 'components/contexts/TopicMessagesContext'; import { useAppSelector } from 'lib/hooks/redux'; +import { Button } from 'components/common/Button/Button'; +import { useSearchParams } from 'react-router-dom'; +import { MESSAGES_PER_PAGE } from 'lib/constants'; +import * as S from 'components/common/NewTable/Table.styled'; import PreviewModal from './PreviewModal'; import Message, { PreviewFilter } from './Message'; @@ -19,75 +23,123 @@ const MessagesTable: React.FC = () => { const [keyFilters, setKeyFilters] = useState<PreviewFilter[]>([]); const [contentFilters, setContentFilters] = useState<PreviewFilter[]>([]); + const [searchParams, setSearchParams] = useSearchParams(); + const page = searchParams.get('page'); const { isLive } = useContext(TopicMessagesContext); const messages = useAppSelector(getTopicMessges); const isFetching = useAppSelector(getIsTopicMessagesFetching); + + const isTailing = isLive && isFetching; + + // Pagination is disabled in live mode, also we don't want to show the button + // if we are fetching the messages or if we are at the end of the topic + const isPaginationDisabled = isTailing || isFetching; + + const isNextPageButtonDisabled = + isPaginationDisabled || messages.length < Number(MESSAGES_PER_PAGE); + const isPrevPageButtonDisabled = + isPaginationDisabled || !Number(searchParams.get('page')); + + const handleNextPage = () => { + searchParams.set('page', String(Number(page || 0) + 1)); + setSearchParams(searchParams); + }; + + const handlePrevPage = () => { + searchParams.set('page', String(Number(page || 0) - 1)); + setSearchParams(searchParams); + }; + return ( - <Table isFullwidth> - <thead> - <tr> - <TableHeaderCell> </TableHeaderCell> - <TableHeaderCell title="Offset" /> - <TableHeaderCell title="Partition" /> - <TableHeaderCell title="Timestamp" /> - <TableHeaderCell - title="Key" - previewText={`Preview ${ - keyFilters.length ? `(${keyFilters.length} selected)` : '' - }`} - onPreview={() => setPreviewFor('key')} - /> - <TableHeaderCell - title="Value" - previewText={`Preview ${ - contentFilters.length ? `(${contentFilters.length} selected)` : '' - }`} - onPreview={() => setPreviewFor('content')} - /> - <TableHeaderCell> </TableHeaderCell> + <> + <Table isFullwidth> + <thead> + <tr> + <TableHeaderCell> </TableHeaderCell> + <TableHeaderCell title="Offset" /> + <TableHeaderCell title="Partition" /> + <TableHeaderCell title="Timestamp" /> + <TableHeaderCell + title="Key" + previewText={`Preview ${ + keyFilters.length ? `(${keyFilters.length} selected)` : '' + }`} + onPreview={() => setPreviewFor('key')} + /> + <TableHeaderCell + title="Value" + previewText={`Preview ${ + contentFilters.length + ? `(${contentFilters.length} selected)` + : '' + }`} + onPreview={() => setPreviewFor('content')} + /> + <TableHeaderCell> </TableHeaderCell> - {previewFor !== null && ( - <PreviewModal - values={previewFor === 'key' ? keyFilters : contentFilters} - toggleIsOpen={() => setPreviewFor(null)} - setFilters={(payload: PreviewFilter[]) => - previewFor === 'key' - ? setKeyFilters(payload) - : setContentFilters(payload) - } + {previewFor !== null && ( + <PreviewModal + values={previewFor === 'key' ? keyFilters : contentFilters} + toggleIsOpen={() => setPreviewFor(null)} + setFilters={(payload: PreviewFilter[]) => + previewFor === 'key' + ? setKeyFilters(payload) + : setContentFilters(payload) + } + /> + )} + </tr> + </thead> + <tbody> + {messages.map((message: TopicMessage) => ( + <Message + key={[ + message.offset, + message.timestamp, + message.key, + message.partition, + ].join('-')} + message={message} + keyFilters={keyFilters} + contentFilters={contentFilters} /> + ))} + {isFetching && isLive && !messages.length && ( + <tr> + <td colSpan={10}> + <PageLoader /> + </td> + </tr> )} - </tr> - </thead> - <tbody> - {messages.map((message: TopicMessage) => ( - <Message - key={[ - message.offset, - message.timestamp, - message.key, - message.partition, - ].join('-')} - message={message} - keyFilters={keyFilters} - contentFilters={contentFilters} - /> - ))} - {isFetching && isLive && !messages.length && ( - <tr> - <td colSpan={10}> - <PageLoader /> - </td> - </tr> - )} - {messages.length === 0 && !isFetching && ( - <tr> - <td colSpan={10}>No messages found</td> - </tr> - )} - </tbody> - </Table> + {messages.length === 0 && !isFetching && ( + <tr> + <td colSpan={10}>No messages found</td> + </tr> + )} + </tbody> + </Table> + <S.Pagination> + <S.Pages> + <Button + buttonType="secondary" + buttonSize="L" + disabled={isPrevPageButtonDisabled} + onClick={handlePrevPage} + > + ← Back + </Button> + <Button + buttonType="secondary" + buttonSize="L" + disabled={isNextPageButtonDisabled} + onClick={handleNextPage} + > + Next β†’ + </Button> + </S.Pages> + </S.Pagination> + </> ); }; diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/__test__/MessagesTable.spec.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Messages/__test__/MessagesTable.spec.tsx index 37ad28b202b..e2aeb68bbba 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/__test__/MessagesTable.spec.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/__test__/MessagesTable.spec.tsx @@ -85,9 +85,10 @@ describe('MessagesTable', () => { }); describe('Custom Setup with different props value', () => { - it('should check if next click is gone during isLive Param', () => { + it('should check if next button and previous is disabled isLive Param', () => { setUpComponent(searchParams, { ...contextValue, isLive: true }); - expect(screen.queryByText(/next/i)).not.toBeInTheDocument(); + expect(screen.queryByText(/next/i)).toBeDisabled(); + expect(screen.queryByText(/back/i)).toBeDisabled(); }); it('should check the display of the loader element', () => {
null
train
val
2022-11-15T13:05:06
"2022-08-25T16:57:13Z"
Haarolean
train
provectus/kafka-ui/2857_2863
provectus/kafka-ui
provectus/kafka-ui/2857
provectus/kafka-ui/2863
[ "keyword_pr_to_issue" ]
7fcbf7507b3ebbd9259b538ba291d76d0f2b8a5f
8bb9f6365823ed8ad181f68ec3a37e6652c4e39e
[]
[ "i think we can use find here instead of iterating over the array twice , \r\n[MDN doc for find](https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Array/find)" ]
"2022-11-02T08:18:17Z"
[ "type/bug", "good first issue", "scope/frontend", "status/accepted", "status/confirmed" ]
Seek type field became empty with refreshing the page
**Describe the bug** Seek type field became empty with refreshing the Topic/Messages **Set up** https://www.kafka-ui.provectus.io/ **Steps to Reproduce** Steps to reproduce the behavior: 1. Navigate to Topic/Messages tab 2. Make sure Seek Type value is "Offset" by default 3. Refresh the page **Expected behavior** The default "Offset" value should stay for Seek Type **Screenshots** https://user-images.githubusercontent.com/104780608/199174719-8ddb7f18-3526-4335-b70c-63c2739d92a5.mov **Additional context** <!-- (Add any other context about the problem here) -->
[ "kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx" ]
[ "kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx index c010ca65cb5..1b99650f715 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx @@ -109,7 +109,11 @@ const Filters: React.FC<FiltersProps> = ({ ); const [currentSeekType, setCurrentSeekType] = React.useState<SeekType>( - (searchParams.get('seekType') as SeekType) || SeekType.OFFSET + SeekTypeOptions.find( + (ele) => ele.value === (searchParams.get('seekType') as SeekType) + ) !== undefined + ? (searchParams.get('seekType') as SeekType) + : SeekType.OFFSET ); const [offset, setOffset] = React.useState<string>( getOffsetFromSeekToParam(searchParams)
null
test
val
2022-11-16T16:36:27
"2022-11-01T06:43:23Z"
armenuikafka
train
provectus/kafka-ui/2699_2864
provectus/kafka-ui
provectus/kafka-ui/2699
provectus/kafka-ui/2864
[ "connected" ]
c4bbbbe3da06779907a57b05354bcfcb9919ce81
b1891c021bc587f7c2ab7b28f13a4745ceab1f49
[]
[ "use var type pls, 'Locator' in this case -> connectorHeaderLocator", "why does it called tabElementLocator? rename to connectorNameLocator for example", "it's not a locator, you already defined it as SelenideElement. seems it's schemaTypeDdl?", "schemaHeaderLocator", "seems here we need to return boolean to check is element by defined locator visible. and also update assertion in regarding test", "seems here we need to return boolean to check is element by defined locator visible. and also update assertion in regarding test", "where do we use this method?", "I have concerns about two different strings with the same value -> //li[@value='%s']", "extra spaces", "we can leave one of them, call it ddlElementLocator and use in both methods", "cleanUpPolicyLocator and partitionsLocator are SelenideElements", "topicHeaderLocator", "no reason to use SoftAssertions to check elements in different places", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "I've asked you to rename tabElementLocator -> connectorNameLocator\nwhy did you create another one element?", "where do we use this method?", "really?", "where do we use this method?", "incomplete, because u find element using its name as locator, then u get in back and compare?", "incomplete, because u find element using its name as locator, then u get in back and compare?", "incomplete, because u find element using its name as locator, then u get in back and compare?", "incomplete, because u find element using its name as locator, then u get in back and compare?", "Removed.", "Removed.", "Renamed and removed.", "Removed.", "Fixed.", "Have done.", "Changed to boolean assert.", "Changed to boolean assert.", "Changed to boolean assert.", "Changed to boolean assert.", "why do we need to apply String.valueOf() to string?", "why do we need to apply String.valueOf() to string?", "here we need to wait until connectorDetails screen loaded", "here we check elements on the same page. why we can't use isSchemaHeaderVisible() check inside SoftAssertions?", "here we check elements on the same page. why we can't use isSchemaHeaderVisible() check inside SoftAssertions?", "here we check elements on the same page. why we can't use isSchemaHeaderVisible() check inside SoftAssertions?", "here we check elements on the same page. why we can't use isSchemaHeaderVisible() check inside SoftAssertions?", "btw, there's no var %s in locator, method is incorrect", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "are u sure in your approach?", "Assertions? inside SoftAssertions? is it Ok??", "Assertions? inside SoftAssertions? is it Ok??", "Assertions? inside SoftAssertions? is it Ok??", "Assertions? inside SoftAssertions? is it Ok??", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "Fixed." ]
"2022-11-02T09:27:04Z"
[ "scope/QA", "scope/AQA" ]
[e2e] review tests to use more object details while creating and updating
we have a lot of missed details (in Topic for example) that we can set while creating and assert than need to review and add as much details as possible and make assertions after creation not only for new object in the list
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Schema.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Topic.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/NaviSideBar.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorCreateForm.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicCreateEditForm.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Schema.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Topic.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/NaviSideBar.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorCreateForm.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicCreateEditForm.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/ConnectorsTests.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/SchemasTests.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Schema.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Schema.java index 8097c2e7d72..a9c7dca75bf 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Schema.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Schema.java @@ -14,19 +14,19 @@ public class Schema { private SchemaType type; public static Schema createSchemaAvro(){ - return new Schema().setName(randomAlphabetic(10)) + return new Schema().setName("schema_avro-" + randomAlphabetic(5)) .setType(SchemaType.AVRO) .setValuePath(System.getProperty("user.dir") + "/src/main/resources/testData/schema_avro_value.json"); } public static Schema createSchemaJson(){ - return new Schema().setName(randomAlphabetic(10)) + return new Schema().setName("schema_json-" + randomAlphabetic(5)) .setType(SchemaType.JSON) .setValuePath(System.getProperty("user.dir") + "/src/main/resources/testData/schema_Json_Value.json"); } public static Schema createSchemaProtobuf(){ - return new Schema().setName(randomAlphabetic(10)) + return new Schema().setName("schema_protobuf-" + randomAlphabetic(5)) .setType(SchemaType.PROTOBUF) .setValuePath(System.getProperty("user.dir") + "/src/main/resources/testData/schema_protobuf_value.txt"); } diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Topic.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Topic.java index 18763e37191..d9a306a4f57 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Topic.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Topic.java @@ -1,10 +1,14 @@ package com.provectus.kafka.ui.models; +import com.provectus.kafka.ui.pages.topic.TopicCreateEditForm; import lombok.Data; import lombok.experimental.Accessors; @Data @Accessors(chain = true) public class Topic { - private String name, cleanupPolicyValue, timeToRetainData, maxSizeOnDisk, maxMessageBytes, messageKey, messageContent ; -} + private String name, timeToRetainData, maxMessageBytes, messageKey, messageContent, + partitions, customParameter; + private TopicCreateEditForm.CleanupPolicyValue cleanupPolicyValue; + private TopicCreateEditForm.MaxSizeOnDisk maxSizeOnDisk; +} \ No newline at end of file diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/NaviSideBar.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/NaviSideBar.java index 0f6f4c22609..565bbd24e5c 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/NaviSideBar.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/NaviSideBar.java @@ -11,6 +11,7 @@ import static com.codeborne.selenide.Selenide.$x; import static com.provectus.kafka.ui.settings.Source.CLUSTER_NAME; +import static com.provectus.kafka.ui.utilities.WebUtils.clickByActions; public class NaviSideBar { @@ -22,7 +23,7 @@ public class NaviSideBar { private SelenideElement expandCluster(String clusterName) { SelenideElement clusterElement = $x(String.format(clusterElementLocator, clusterName)).shouldBe(Condition.visible); if (clusterElement.parent().$$x(".//ul").size() == 0) { - clusterElement.click(); + clickByActions(clusterElement); } return clusterElement; } @@ -36,10 +37,8 @@ public NaviSideBar waitUntilScreenReady() { @Step public NaviSideBar openSideMenu(String clusterName, SideMenuOption option) { - expandCluster(clusterName) - .parent() - .$x(String.format(sideMenuOptionElementLocator, option.value)) - .click(); + clickByActions(expandCluster(clusterName).parent() + .$x(String.format(sideMenuOptionElementLocator, option.value))); return this; } diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorCreateForm.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorCreateForm.java index 5c007137465..26cab86f63b 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorCreateForm.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorCreateForm.java @@ -5,6 +5,7 @@ import io.qameta.allure.Step; import static com.codeborne.selenide.Selenide.$x; +import static com.codeborne.selenide.Selenide.refresh; import static com.provectus.kafka.ui.utilities.WebUtils.clickByJavaScript; public class ConnectorCreateForm { @@ -23,13 +24,18 @@ public ConnectorCreateForm waitUntilScreenReady() { } @Step - public ConnectorCreateForm setConnectorConfig(String connectName, String configJson) { + public ConnectorCreateForm setConnectorDetails(String connectName, String configJson) { nameField.shouldBe(Condition.enabled).setValue(connectName); configField.shouldBe(Condition.enabled).click(); contentTextArea.setValue(configJson); nameField.shouldBe(Condition.enabled).click(); - clickByJavaScript(submitBtn); - loadingSpinner.shouldBe(Condition.disappear); return this; } -} \ No newline at end of file + + @Step + public ConnectorCreateForm clickSubmitButton() { + clickByJavaScript(submitBtn); + loadingSpinner.shouldBe(Condition.disappear); + return this; + } +} diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorDetails.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorDetails.java index c099c60563d..e5e6b4cbd3c 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorDetails.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/ConnectorDetails.java @@ -1,35 +1,36 @@ package com.provectus.kafka.ui.pages.connector; +import static com.codeborne.selenide.Selenide.$x; +import static com.provectus.kafka.ui.utilities.WebUtils.clearByKeyboard; +import static com.provectus.kafka.ui.utilities.WebUtils.clickByJavaScript; +import static com.provectus.kafka.ui.utilities.WebUtils.isVisible; + import com.codeborne.selenide.Condition; import com.codeborne.selenide.SelenideElement; import io.qameta.allure.Step; - +import java.time.Duration; import java.util.Arrays; -import static com.codeborne.selenide.Selenide.$x; -import static com.provectus.kafka.ui.utilities.WebUtils.clearByKeyboard; -import static com.provectus.kafka.ui.utilities.WebUtils.clickByJavaScript; - public class ConnectorDetails { - protected SelenideElement loadingSpinner = $x("//*[contains(text(),'Loading')]"); - protected SelenideElement dotMenuBtn = $x("//button[@aria-label='Dropdown Toggle']"); - protected SelenideElement deleteBtn = $x("//li/div[contains(text(),'Delete')]"); - protected SelenideElement confirmBtnMdl = $x("//div[@role='dialog']//button[contains(text(),'Confirm')]"); - protected SelenideElement submitBtn = $x("//button[@type='submit']"); - protected SelenideElement contentTextArea = $x("//textarea[@class='ace_text-input']"); - protected SelenideElement taskTab = $x("//a[contains(text(),'Tasks')]"); - protected SelenideElement configTab = $x("//a[contains(text(),'Config')]"); - protected SelenideElement configField = $x("//div[@id='config']"); - protected SelenideElement successAlertMessage = $x("//div[contains(text(),'Config successfully updated')]"); + protected SelenideElement loadingSpinner = $x("//*[contains(text(),'Loading')]"); + protected SelenideElement dotMenuBtn = $x("//button[@aria-label='Dropdown Toggle']"); + protected SelenideElement deleteBtn = $x("//li/div[contains(text(),'Delete')]"); + protected SelenideElement confirmBtnMdl = $x("//div[@role='dialog']//button[contains(text(),'Confirm')]"); + protected SelenideElement submitBtn = $x("//button[@type='submit']"); + protected SelenideElement contentTextArea = $x("//textarea[@class='ace_text-input']"); + protected SelenideElement taskTab = $x("//a[contains(text(),'Tasks')]"); + protected SelenideElement configTab = $x("//a[contains(text(),'Config')]"); + protected SelenideElement configField = $x("//div[@id='config']"); + protected SelenideElement successAlertMessage = $x("//div[contains(text(),'Config successfully updated')]"); + protected String connectorHeaderLocator = "//h1[contains(text(),'%s')]"; - - @Step - public ConnectorDetails waitUntilScreenReady() { - loadingSpinner.shouldBe(Condition.disappear); - Arrays.asList(taskTab,configTab).forEach(elementsMenu -> elementsMenu.shouldBe(Condition.visible)); - return this; - } + @Step + public ConnectorDetails waitUntilScreenReady() { + loadingSpinner.shouldBe(Condition.disappear); + dotMenuBtn.shouldBe(Condition.visible); + return this; + } @Step public ConnectorDetails openConfigTab() { @@ -43,11 +44,16 @@ public ConnectorDetails setConfig(String configJson) { clearByKeyboard(contentTextArea); contentTextArea.setValue(configJson); configField.shouldBe(Condition.enabled).click(); - clickByJavaScript(submitBtn); - successAlertMessage.shouldBe(Condition.visible); return this; } + @Step + public ConnectorDetails clickSubmitButton() { + clickByJavaScript(submitBtn); + successAlertMessage.shouldBe(Condition.visible); + return this; + } + @Step public ConnectorDetails openDotMenu() { clickByJavaScript(dotMenuBtn); @@ -74,4 +80,9 @@ public ConnectorDetails deleteConnector() { clickConfirmBtn(); return this; } -} \ No newline at end of file + + @Step + public boolean isConnectorHeaderVisible(String connectorName) { + return isVisible($x(String.format(connectorHeaderLocator,connectorName))); + } +} diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java index 36a905d2d91..16104ac1ca8 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java @@ -15,16 +15,15 @@ public class KafkaConnectList { protected SelenideElement loadingSpinner = $x("//*[contains(text(),'Loading')]"); - protected SelenideElement pageTitle = $x("//h1[text()='Connectors']"); protected SelenideElement createConnectorBtn = $x("//button[contains(text(),'Create Connector')]"); protected SelenideElement connectorsGrid = $x("//table"); - protected String tabElementLocator = "//td[contains(text(),'%s')]"; + protected String connectorNameLocator = "//td[contains(text(),'%s')]"; @Step public KafkaConnectList waitUntilScreenReady() { - loadingSpinner.shouldBe(Condition.disappear); - pageTitle.shouldBe(Condition.visible); - return this; + loadingSpinner.shouldBe(Condition.disappear); + createConnectorBtn.shouldBe(Condition.visible); + return this; } @Step @@ -35,7 +34,7 @@ public KafkaConnectList clickCreateConnectorBtn() { @Step public KafkaConnectList openConnector(String connectorName) { - $x(String.format(tabElementLocator,connectorName)) + $x(String.format(connectorNameLocator, connectorName)) .shouldBe(Condition.enabled).click(); return this; } @@ -43,13 +42,13 @@ public KafkaConnectList openConnector(String connectorName) { @Step public boolean isConnectorVisible(String connectorName) { connectorsGrid.shouldBe(Condition.visible); - return isVisible($x(String.format(tabElementLocator,connectorName))); + return isVisible($x(String.format(connectorNameLocator,connectorName))); } @Step public KafkaConnectList connectorIsUpdatedInList(String connectorName, String topicName) { - $x(String.format(tabElementLocator,connectorName)).shouldBe(Condition.visible); - By.xpath(String.format(tabElementLocator,topicName)).refreshUntil(Condition.visible); + $x(String.format(connectorNameLocator,connectorName)).shouldBe(Condition.visible); + By.xpath(String.format(connectorNameLocator,topicName)).refreshUntil(Condition.visible); return this; } } diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaDetails.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaDetails.java index e3f392514b5..981de748e99 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaDetails.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaDetails.java @@ -7,6 +7,7 @@ import static com.codeborne.selenide.Selenide.$$x; import static com.codeborne.selenide.Selenide.$x; import static com.provectus.kafka.ui.utilities.WebUtils.clickByJavaScript; +import static com.provectus.kafka.ui.utilities.WebUtils.isVisible; public class SchemaDetails { @@ -17,6 +18,8 @@ public class SchemaDetails { protected SelenideElement editSchemaBtn = $x("//button[contains(text(),'Edit Schema')]"); protected SelenideElement removeBtn = $x("//*[contains(text(),'Remove')]"); protected SelenideElement confirmBtn = $x("//div[@role='dialog']//button[contains(text(),'Confirm')]"); + protected SelenideElement schemaTypeDdl = $x("//h4[contains(text(),'Type')]/../p"); + protected String schemaHeaderLocator = "//h1[contains(text(),'%s')]"; @Step public SchemaDetails waitUntilScreenReady() { @@ -30,6 +33,16 @@ public String getCompatibility() { return compatibilityField.getText(); } + @Step + public boolean isSchemaHeaderVisible(String schemaName) { + return isVisible($x(String.format(schemaHeaderLocator,schemaName))); + } + + @Step + public String getSchemaType() { + return schemaTypeDdl.getText(); + } + @Step public SchemaDetails openEditSchema(){ editSchemaBtn.shouldBe(Condition.visible).click(); diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicCreateEditForm.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicCreateEditForm.java index 8a71ddb6ffb..539520445ed 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicCreateEditForm.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicCreateEditForm.java @@ -14,12 +14,14 @@ public class TopicCreateEditForm { protected SelenideElement loadingSpinner = $x("//*[contains(text(),'Loading')]"); protected SelenideElement timeToRetainField = $x("//input[@id='timeToRetain']"); + protected SelenideElement partitionsField = $x("//input[@name='partitions']"); protected SelenideElement nameField = $x("//input[@name='name']"); protected SelenideElement maxMessageBytesField = $x("//input[@name='maxMessageBytes']"); protected SelenideElement minInSyncReplicasField = $x("//input[@name='minInSyncReplicas']"); protected SelenideElement cleanUpPolicyDdl = $x("//ul[@id='topicFormCleanupPolicy']"); + protected SelenideElement maxSizeOnDiscDdl = $x("//ul[@id='topicFormRetentionBytes']"); protected SelenideElement createTopicBtn = $x("//button[@type='submit']"); - protected String cleanUpPolicyTypeLocator = "//li[text()='%s']"; + protected String ddlElementLocator = "//li[@value='%s']"; @Step public TopicCreateEditForm waitUntilScreenReady(){ @@ -53,8 +55,9 @@ public TopicCreateEditForm setTimeToRetainDataInMs(String ms) { } @Step - public TopicCreateEditForm setMaxSizeOnDiskInGB(String value) { - new KafkaUISelectElement("retentionBytes").selectByVisibleText(value); + public TopicCreateEditForm setMaxSizeOnDiskInGB(MaxSizeOnDisk MaxSizeOnDisk) { + maxSizeOnDiscDdl.shouldBe(Condition.visible).click(); + $x(String.format(ddlElementLocator, MaxSizeOnDisk.getOptionValue())).shouldBe(Condition.visible).click(); return this; } @@ -69,6 +72,12 @@ public TopicCreateEditForm setMaxMessageBytes(String bytes) { return setMaxMessageBytes(Long.parseLong(bytes)); } + @Step + public TopicCreateEditForm setPartitions(String partitions){ + partitionsField.setValue(partitions); + return this; + } + @Step public TopicCreateEditForm setTimeToRetainDataInMsUsingButtons(String value) { timeToRetainField @@ -81,15 +90,9 @@ public TopicCreateEditForm setTimeToRetainDataInMsUsingButtons(String value) { } @Step - public TopicCreateEditForm selectCleanupPolicy(CleanupPolicyValue cleanupPolicyValue) { - return selectFromDropDownByOptionValue("cleanupPolicy", - cleanupPolicyValue.getOptionValue()); - } - - @Step - public TopicCreateEditForm selectCleanupPolicy(String cleanupPolicyOptionValue) { + public TopicCreateEditForm selectCleanupPolicy(CleanupPolicyValue cleanupPolicyOptionValue) { cleanUpPolicyDdl.shouldBe(Condition.visible).click(); - $x(String.format(cleanUpPolicyTypeLocator,cleanupPolicyOptionValue)).shouldBe(Condition.visible).click(); + $x(String.format(ddlElementLocator,cleanupPolicyOptionValue.getOptionValue())).shouldBe(Condition.visible).click(); return this; } @@ -242,6 +245,30 @@ public String getVisibleText() { } } + public enum MaxSizeOnDisk { + NOT_SET("-1", "Not Set"), + SIZE_1_GB("1073741824", "1 GB"), + SIZE_10_GB("10737418240", "10 GB"), + SIZE_20_GB("21474836480", "20 GB"), + SIZE_50_GB("53687091200", "50 GB"); + + private final String optionValue; + private final String visibleText; + + MaxSizeOnDisk(String optionValue, String visibleText) { + this.optionValue = optionValue; + this.visibleText = visibleText; + } + + public String getOptionValue() { + return optionValue; + } + + public String getVisibleText() { + return visibleText; + } + } + private TopicCreateEditForm selectFromDropDownByOptionValue(String dropDownElementName, String optionValue) { KafkaUISelectElement select = new KafkaUISelectElement(dropDownElementName); diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java index 73ce320a485..a065169c0ce 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java @@ -11,6 +11,7 @@ import static com.codeborne.selenide.Selenide.*; import static com.provectus.kafka.ui.utilities.WebUtils.clickByJavaScript; +import static com.provectus.kafka.ui.utilities.WebUtils.isVisible; @ExtensionMethod({WaitUtils.class}) public class TopicDetails { @@ -26,7 +27,10 @@ public class TopicDetails { protected SelenideElement confirmBtn = $x("//div[@role='dialog']//button[contains(text(),'Confirm')]"); protected SelenideElement produceMessageBtn = $x("//div//button[text()='Produce Message']"); protected SelenideElement contentMessageTab = $x("//html//div[@id='root']/div/main//table//p"); + protected SelenideElement cleanUpPolicyField = $x("//div[contains(text(),'Clean Up Policy')]/../span/*"); + protected SelenideElement partitionsField = $x("//div[contains(text(),'Partitions')]/../span"); protected String consumerIdLocator = "//a[@title='%s']"; + protected String topicHeaderLocator = "//h1[contains(text(),'%s')]"; @Step public TopicDetails waitUntilScreenReady() { @@ -54,6 +58,21 @@ public TopicDetails openDotPartitionIdMenu() { return this; } + @Step + public String getCleanUpPolicy(){ + return cleanUpPolicyField.getText(); + } + + @Step + public String getPartitions(){ + return partitionsField.getText(); + } + + @Step + public boolean isTopicHeaderVisible(String topicName) { + return isVisible($x(String.format(topicHeaderLocator,topicName))); + } + @Step public TopicDetails clickClearMessagesBtn() { clearMessagesBtn.shouldBe(Condition.visible.because("Clear Messages invisible")).click(); @@ -74,7 +93,6 @@ public TopicDetails clickProduceMessageBtn() { clickByJavaScript(produceMessageBtn); return this; } - @Step public TopicDetails openConsumerGroup(String consumerId) { $x(String.format(consumerIdLocator, consumerId)).click(); @@ -158,4 +176,4 @@ public String toString() { return "TopicMenu{" + "value='" + value + '\'' + '}'; } } -} \ No newline at end of file +} diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java index 7215cb7b86b..ad93bfbc772 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java @@ -2,14 +2,25 @@ import com.codeborne.selenide.Condition; import com.codeborne.selenide.SelenideElement; +import com.codeborne.selenide.WebDriverRunner; import lombok.extern.slf4j.Slf4j; import org.openqa.selenium.Keys; +import org.openqa.selenium.interactions.Actions; import static com.codeborne.selenide.Selenide.executeJavaScript; @Slf4j public class WebUtils { + public static void clickByActions(SelenideElement element) { + log.debug("\nclickByActions: {}", element.getSearchCriteria()); + element.shouldBe(Condition.enabled); + new Actions(WebDriverRunner.getWebDriver()) + .moveToElement(element) + .click(element) + .perform(); + } + public static void clickByJavaScript(SelenideElement element) { log.debug("\nclickByJavaScript: {}", element.getSearchCriteria()); element.shouldBe(Condition.enabled); @@ -34,4 +45,5 @@ public static boolean isVisible(SelenideElement element) { } return isVisible; } + } \ No newline at end of file
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/ConnectorsTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/ConnectorsTests.java index bd06882887f..1203fc9ebcf 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/ConnectorsTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/ConnectorsTests.java @@ -15,6 +15,7 @@ import static com.provectus.kafka.ui.pages.NaviSideBar.SideMenuOption.KAFKA_CONNECT; import static com.provectus.kafka.ui.settings.Source.CLUSTER_NAME; import static com.provectus.kafka.ui.utilities.FileUtils.getResourceAsString; +import static org.apache.commons.lang.RandomStringUtils.randomAlphabetic; @TestInstance(TestInstance.Lifecycle.PER_CLASS) public class ConnectorsTests extends BaseTest { @@ -26,19 +27,19 @@ public class ConnectorsTests extends BaseTest { private static final String MESSAGE_CONTENT = "message_content_create_topic.json"; private static final String MESSAGE_KEY = " "; private static final Topic TOPIC_FOR_CREATE = new Topic() - .setName("topic_for_create_connector") + .setName("topic_for_create_connector-" + randomAlphabetic(5)) .setMessageContent(MESSAGE_CONTENT).setMessageKey(MESSAGE_KEY); private static final Topic TOPIC_FOR_DELETE = new Topic() - .setName("topic_for_delete_connector") + .setName("topic_for_delete_connector-" + randomAlphabetic(5)) .setMessageContent(MESSAGE_CONTENT).setMessageKey(MESSAGE_KEY); private static final Topic TOPIC_FOR_UPDATE = new Topic() - .setName("topic_for_update_connector") + .setName("topic_for_update_connector-" + randomAlphabetic(5)) .setMessageContent(MESSAGE_CONTENT).setMessageKey(MESSAGE_KEY); private static final Connector CONNECTOR_FOR_DELETE = new Connector() - .setName("sink_postgres_activities_e2e_checks_for_delete") + .setName("sink_postgres_activities_e2e_checks_for_delete-" + randomAlphabetic(5)) .setConfig(getResourceAsString("delete_connector_config.json")); private static final Connector CONNECTOR_FOR_UPDATE = new Connector() - .setName("sink_postgres_activities_e2e_checks_for_update") + .setName("sink_postgres_activities_e2e_checks_for_update-" + randomAlphabetic(5)) .setConfig(getResourceAsString("config_for_create_connector_via_api.json")); @BeforeAll @@ -60,7 +61,7 @@ public void beforeAll() { @Test public void createConnector() { Connector connectorForCreate = new Connector() - .setName("sink_postgres_activities_e2e_checks") + .setName("sink_postgres_activities_e2e_checks-" + randomAlphabetic(5)) .setConfig(getResourceAsString("config_for_create_connector.json")); naviSideBar .openSideMenu(KAFKA_CONNECT); @@ -69,12 +70,23 @@ public void createConnector() { .clickCreateConnectorBtn(); connectorCreateForm .waitUntilScreenReady() - .setConnectorConfig(connectorForCreate.getName(), connectorForCreate.getConfig()); + .setConnectorDetails(connectorForCreate.getName(), connectorForCreate.getConfig()) + .clickSubmitButton(); + connectorDetails + .waitUntilScreenReady(); + naviSideBar + .openSideMenu(KAFKA_CONNECT); + kafkaConnectList + .waitUntilScreenReady() + .openConnector(connectorForCreate.getName()); + connectorDetails + .waitUntilScreenReady(); + Assertions.assertTrue(connectorDetails.isConnectorHeaderVisible(connectorForCreate.getName()),"isConnectorTitleVisible()"); naviSideBar .openSideMenu(KAFKA_CONNECT); kafkaConnectList .waitUntilScreenReady(); - Assertions.assertTrue(kafkaConnectList.isConnectorVisible(connectorForCreate.getName()), "isConnectorVisible()"); + Assertions.assertTrue(kafkaConnectList.isConnectorVisible(CONNECTOR_FOR_DELETE.getName()), "isConnectorVisible()"); CONNECTOR_LIST.add(connectorForCreate); } @@ -92,7 +104,8 @@ public void updateConnector() { connectorDetails .waitUntilScreenReady() .openConfigTab() - .setConfig(CONNECTOR_FOR_UPDATE.getConfig()); + .setConfig(CONNECTOR_FOR_UPDATE.getConfig()) + .clickSubmitButton(); naviSideBar .openSideMenu(KAFKA_CONNECT); kafkaConnectList diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/SchemasTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/SchemasTests.java index e76f28728e1..e0c33e2f3c0 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/SchemasTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/SchemasTests.java @@ -8,6 +8,7 @@ import com.provectus.kafka.ui.utilities.qaseIoUtils.enums.Status; import io.qase.api.annotation.CaseId; import lombok.SneakyThrows; +import org.assertj.core.api.SoftAssertions; import org.junit.jupiter.api.*; import java.util.ArrayList; @@ -54,11 +55,16 @@ void createSchemaAvro() { .clickSubmitBtn(); schemaDetails .waitUntilScreenReady(); + SoftAssertions softly = new SoftAssertions(); + softly.assertThat(schemaDetails.isSchemaHeaderVisible(schemaAvro.getName())).as("isSchemaHeaderVisible()").isTrue(); + softly.assertThat(schemaDetails.getSchemaType()).as("getSchemaType()").isEqualTo(schemaAvro.getType().getValue()); + softly.assertThat(schemaDetails.getCompatibility()).as("getCompatibility()").isEqualTo(CompatibilityLevel.CompatibilityEnum.BACKWARD.getValue()); + softly.assertAll(); naviSideBar .openSideMenu(SCHEMA_REGISTRY); schemaRegistryList .waitUntilScreenReady(); - Assertions.assertTrue(schemaRegistryList.isSchemaVisible(schemaAvro.getName()),"isSchemaVisible()"); + Assertions.assertTrue(schemaRegistryList.isSchemaVisible(AVRO_API.getName()),"isSchemaVisible()"); SCHEMA_LIST.add(schemaAvro); } @@ -131,11 +137,16 @@ void createSchemaJson() { .clickSubmitBtn(); schemaDetails .waitUntilScreenReady(); + SoftAssertions softly = new SoftAssertions(); + softly.assertThat(schemaDetails.isSchemaHeaderVisible(schemaJson.getName())).as("isSchemaHeaderVisible()").isTrue(); + softly.assertThat(schemaDetails.getSchemaType()).as("getSchemaType()").isEqualTo(schemaJson.getType().getValue()); + softly.assertThat(schemaDetails.getCompatibility()).as("getCompatibility()").isEqualTo(CompatibilityLevel.CompatibilityEnum.BACKWARD.getValue()); + softly.assertAll(); naviSideBar .openSideMenu(SCHEMA_REGISTRY); schemaRegistryList .waitUntilScreenReady(); - Assertions.assertTrue(schemaRegistryList.isSchemaVisible(schemaJson.getName()),"isSchemaVisible()"); + Assertions.assertTrue(schemaRegistryList.isSchemaVisible(JSON_API.getName()),"isSchemaVisible()"); SCHEMA_LIST.add(schemaJson); } @@ -180,11 +191,16 @@ void createSchemaProtobuf() { .clickSubmitBtn(); schemaDetails .waitUntilScreenReady(); + SoftAssertions softly = new SoftAssertions(); + softly.assertThat(schemaDetails.isSchemaHeaderVisible(schemaProtobuf.getName())).as("isSchemaHeaderVisible()").isTrue(); + softly.assertThat(schemaDetails.getSchemaType()).as("getSchemaType()").isEqualTo(schemaProtobuf.getType().getValue()); + softly.assertThat(schemaDetails.getCompatibility()).as("getCompatibility()").isEqualTo(CompatibilityLevel.CompatibilityEnum.BACKWARD.getValue()); + softly.assertAll(); naviSideBar .openSideMenu(SCHEMA_REGISTRY); schemaRegistryList .waitUntilScreenReady(); - Assertions.assertTrue(schemaRegistryList.isSchemaVisible(schemaProtobuf.getName()),"isSchemaVisible()"); + Assertions.assertTrue(schemaRegistryList.isSchemaVisible(PROTOBUF_API.getName()),"isSchemaVisible()"); SCHEMA_LIST.add(schemaProtobuf); } diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java index 33ad3d015a1..97f70c194e3 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java @@ -15,29 +15,36 @@ import java.util.List; import static com.provectus.kafka.ui.pages.NaviSideBar.SideMenuOption.TOPICS; +import static com.provectus.kafka.ui.pages.topic.TopicCreateEditForm.CleanupPolicyValue.COMPACT; +import static com.provectus.kafka.ui.pages.topic.TopicCreateEditForm.CleanupPolicyValue.DELETE; +import static com.provectus.kafka.ui.pages.topic.TopicCreateEditForm.MaxSizeOnDisk.SIZE_20_GB; import static com.provectus.kafka.ui.settings.Source.CLUSTER_NAME; import static com.provectus.kafka.ui.utilities.FileUtils.fileToString; +import static org.apache.commons.lang.RandomStringUtils.randomAlphabetic; import static org.assertj.core.api.Assertions.assertThat; @TestInstance(TestInstance.Lifecycle.PER_CLASS) public class TopicTests extends BaseTest { private static final long SUITE_ID = 2; private static final String SUITE_TITLE = "Topics"; - + private static final Topic TOPIC_TO_CREATE = new Topic() + .setName("new-topic-"+ randomAlphabetic(5)) + .setPartitions("1") + .setCleanupPolicyValue(DELETE); private static final Topic TOPIC_FOR_UPDATE = new Topic() - .setName("topic-to-update") - .setCleanupPolicyValue("Compact") + .setName("topic-to-update-" + randomAlphabetic(5)) + .setCleanupPolicyValue(COMPACT) .setTimeToRetainData("604800001") - .setMaxSizeOnDisk("20 GB") + .setMaxSizeOnDisk(SIZE_20_GB) .setMaxMessageBytes("1000020") .setMessageKey(fileToString(System.getProperty("user.dir") + "/src/test/resources/producedkey.txt")) .setMessageContent(fileToString(System.getProperty("user.dir") + "/src/test/resources/testData.txt")); private static final Topic TOPIC_FOR_MESSAGES = new Topic() - .setName("topic-with-clean-message-attribute") + .setName("topic-with-clean-message-attribute-" + randomAlphabetic(5)) .setMessageKey(fileToString(System.getProperty("user.dir") + "/src/test/resources/producedkey.txt")) .setMessageContent(fileToString(System.getProperty("user.dir") + "/src/test/resources/testData.txt")); - private static final Topic TOPIC_FOR_DELETE = new Topic().setName("topic-to-delete"); + private static final Topic TOPIC_FOR_DELETE = new Topic().setName("topic-to-delete-" + randomAlphabetic(5)); private static final List<Topic> TOPIC_LIST = new ArrayList<>(); @BeforeAll @@ -52,7 +59,6 @@ public void beforeAll() { @CaseId(199) @Test public void createTopic() { - Topic topicToCreate = new Topic().setName("new-topic"); naviSideBar .openSideMenu(TOPICS); topicsList @@ -60,16 +66,28 @@ public void createTopic() { .clickAddTopicBtn(); topicCreateEditForm .waitUntilScreenReady() - .setTopicName(topicToCreate.getName()) + .setTopicName(TOPIC_TO_CREATE.getName()) + .setPartitions(TOPIC_TO_CREATE.getPartitions()) + .selectCleanupPolicy(TOPIC_TO_CREATE.getCleanupPolicyValue()) .clickCreateTopicBtn(); topicDetails .waitUntilScreenReady(); + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .openTopic(TOPIC_TO_CREATE.getName()); + SoftAssertions softly = new SoftAssertions(); + softly.assertThat(topicDetails.isTopicHeaderVisible(TOPIC_TO_CREATE.getName())).as("isTopicHeaderVisible()").isTrue(); + softly.assertThat(topicDetails.getCleanUpPolicy()).as("getCleanUpPolicy()").isEqualTo(TOPIC_TO_CREATE.getCleanupPolicyValue().toString()); + softly.assertThat(topicDetails.getPartitions()).as("getPartitions()").isEqualTo(TOPIC_TO_CREATE.getPartitions()); + softly.assertAll(); naviSideBar .openSideMenu(TOPICS); topicsList .waitUntilScreenReady(); - Assertions.assertTrue(topicsList.isTopicVisible(topicToCreate.getName()), "isTopicVisible"); - TOPIC_LIST.add(topicToCreate); + Assertions.assertTrue(topicsList.isTopicVisible(TOPIC_TO_CREATE.getName()), "isTopicVisible"); + TOPIC_LIST.add(TOPIC_TO_CREATE); } @Disabled("https://github.com/provectus/kafka-ui/issues/2625") @@ -89,7 +107,7 @@ public void updateTopic() { .openEditSettings(); topicCreateEditForm .waitUntilScreenReady() - .selectCleanupPolicy(TOPIC_FOR_UPDATE.getCleanupPolicyValue()) + .selectCleanupPolicy((TOPIC_FOR_UPDATE.getCleanupPolicyValue())) .setMinInsyncReplicas(10) .setTimeToRetainDataInMs(TOPIC_FOR_UPDATE.getTimeToRetainData()) .setMaxSizeOnDiskInGB(TOPIC_FOR_UPDATE.getMaxSizeOnDisk()) @@ -106,10 +124,10 @@ public void updateTopic() { .waitUntilScreenReady() .openEditSettings(); SoftAssertions softly = new SoftAssertions(); - softly.assertThat(topicCreateEditForm.getCleanupPolicy()).as("Cleanup Policy").isEqualTo(TOPIC_FOR_UPDATE.getCleanupPolicyValue()); - softly.assertThat(topicCreateEditForm.getTimeToRetain()).as("Time to retain").isEqualTo(TOPIC_FOR_UPDATE.getTimeToRetainData()); - softly.assertThat(topicCreateEditForm.getMaxSizeOnDisk()).as("Max size on disk").isEqualTo(TOPIC_FOR_UPDATE.getMaxSizeOnDisk()); - softly.assertThat(topicCreateEditForm.getMaxMessageBytes()).as("Max message bytes").isEqualTo(TOPIC_FOR_UPDATE.getMaxMessageBytes()); + softly.assertThat(topicCreateEditForm.getCleanupPolicy()).as("getCleanupPolicy()").isEqualTo(TOPIC_FOR_UPDATE.getCleanupPolicyValue().getVisibleText()); + softly.assertThat(topicCreateEditForm.getTimeToRetain()).as("getTimeToRetain()").isEqualTo(TOPIC_FOR_UPDATE.getTimeToRetainData()); + softly.assertThat(topicCreateEditForm.getMaxSizeOnDisk()).as("getMaxSizeOnDisk()").isEqualTo(TOPIC_FOR_UPDATE.getMaxSizeOnDisk().getVisibleText()); + softly.assertThat(topicCreateEditForm.getMaxMessageBytes()).as("getMaxMessageBytes()").isEqualTo(TOPIC_FOR_UPDATE.getMaxMessageBytes()); softly.assertAll(); } @@ -134,7 +152,7 @@ public void deleteTopic() { Assertions.assertFalse(topicsList.isTopicVisible(TOPIC_FOR_DELETE.getName()), "isTopicVisible"); TOPIC_LIST.remove(TOPIC_FOR_DELETE); } - + @DisplayName("produce message") @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) @AutomationStatus(status = Status.AUTOMATED)
train
val
2022-11-03T18:16:04
"2022-10-04T06:31:26Z"
VladSenyuta
train
provectus/kafka-ui/2669_2873
provectus/kafka-ui
provectus/kafka-ui/2669
provectus/kafka-ui/2873
[ "connected" ]
5900f8e957f2abf078bb6db4552b186a9e21a0b5
7fcbf7507b3ebbd9259b538ba291d76d0f2b8a5f
[ "Hello @nitin1677 , thank you for issue.\r\n\r\ncan you please share `foreignExchangeRate` field definition (like you have in avro schema) ?\r\n", "{\r\n\t\t\t\t\t\t\"name\": \"foreignExchangeRate\",\r\n\t\t\t\t\t\t\"type\": [\r\n\t\t\t\t\t\t\t\"null\",\r\n\t\t\t\t\t\t\t{\r\n\t\t\t\t\t\t\t\t\"type\": \"bytes\",\r\n\t\t\t\t\t\t\t\t\"logicalType\": \"decimal\",\r\n\t\t\t\t\t\t\t\t\"precision\": 24,\r\n\t\t\t\t\t\t\t\t\"scale\": 16\r\n\t\t\t\t\t\t\t}\r\n\t\t\t\t\t\t],\r\n\t\t\t\t\t\t\"doc\": \"foreign exchange rate\",\r\n\t\t\t\t\t\t\"default\": null\r\n\t\t\t\t\t}", "@nitin1677 , thank you for issue. \r\n\r\nThere are two parts:\r\n1. BE: currently kafka-ui doesnt support logical coversions (I created ticket for that, since it requires some time to implement https://github.com/provectus/kafka-ui/issues/2764) \r\n2. FE: I guess there is some error with downloading `#οΏ½Γ²oÁ\\u0000\\u0000` symbols - lets check and fix that anyway\r\n\r\ncc @Haarolean ", "@iliax can we get an example schema/topic on our dev env for our frontend engineers to reproduce it easily?", "@Haarolean I created topic `issue-2669` and submitted message that reproduces this issue", "The issue has been closed but still this issue exists from latest pull of master.\r\n\r\n\t\t\"field1\": {\r\n\t\t\t\"bytes\": \"#†òoÁ\\\\u0000\\\\u0000\"\r\n\t\t},\r\n\r\n\t\t\"field2\": {\r\n\t\t\t\"bytes\": \"#†òoÁ\\\\u0000\\\\u0000\"\r\n\t\t},\r\n\r\n\t\t\"field3\": {\r\n\t\t\t\"com.avro.schema.common.BigAmount_30_10\": {\r\n\t\t\t\t\"amount\": \"\\\\u0013ΒˆΓ”g\\\\u0000\",\r\n\t\t\t\t\"currency\": \"USD\"\r\n\t\t\t}\r\n\t\t},\r\n\r\n", "@nitin1677 which commit are you on? left top corner in ui", "@nitin1677 please note, that within this issue we only fixed file exportion. Logical type support is not implemented yet (https://github.com/provectus/kafka-ui/issues/2764)" ]
[]
"2022-11-03T13:14:27Z"
[ "type/bug", "good first issue", "scope/frontend", "status/accepted" ]
Save to file issues with bigdecimal
<!-- Don't forget to check for existing issues/discussions regarding your proposal. We might already have it. https://github.com/provectus/kafka-ui/issues https://github.com/provectus/kafka-ui/discussions --> **Describe the bug** <!--(A clear and concise description of what the bug is.)--> when payload contains bigdecimal values, values are displayed as unicode characters like, "foreignExchangeRate": { "bytes": "#†òoÁ\u0000\u0000" } When we select "save as a file", the result in file is truncated after bytes string as no text after this string in output file. ,\"foreignExchangeRate\":{\"bytes\":\" **Set up** <!-- How do you run the app? Please provide as much info as possible: 1. App version (docker image version or check commit hash in the top left corner in UI) 2. Helm chart version, if you use one 3. Any IAAC configs We might close the issue without further explanation if you don't provide such information. --> d149d26013a81e22e5d0c8b59db56ac0565b9e55 https://github.com/provectus/kafka-ui/commit/d149d26 **Steps to Reproduce** <!-- We'd like you to provide an example setup (via docker-compose, helm, etc.) to reproduce the problem, especially with a complex setups. --> Steps to reproduce the behavior: 1. **Expected behavior** <!-- (A clear and concise description of what you expected to happen) --> expected value to be displayed as "foreignExchangeRate": 1.0000000000000000 instead of "foreignExchangeRate": { "bytes": "#†òoÁ\u0000\u0000" } and payload should be downloaded as complete. Note: with latest tag (v0.4.0), even the values are displayed as multi-byte, but save as a file copies the complete payload to file. **Screenshots** <!-- (If applicable, add screenshots to help explain your problem) --> **Additional context** <!-- (Add any other context about the problem here) -->
[ "kafka-ui-react-app/src/lib/hooks/__tests__/useDataSaver.spec.tsx", "kafka-ui-react-app/src/lib/hooks/useDataSaver.ts" ]
[ "kafka-ui-react-app/src/lib/hooks/__tests__/useDataSaver.spec.tsx", "kafka-ui-react-app/src/lib/hooks/useDataSaver.ts" ]
[]
diff --git a/kafka-ui-react-app/src/lib/hooks/__tests__/useDataSaver.spec.tsx b/kafka-ui-react-app/src/lib/hooks/__tests__/useDataSaver.spec.tsx index 510fb955455..f0631850e5c 100644 --- a/kafka-ui-react-app/src/lib/hooks/__tests__/useDataSaver.spec.tsx +++ b/kafka-ui-react-app/src/lib/hooks/__tests__/useDataSaver.spec.tsx @@ -15,30 +15,8 @@ describe('useDataSaver hook', () => { afterAll(() => jest.useRealTimers()); - it('downloads json file', () => { - const link: HTMLAnchorElement = document.createElement('a'); - link.click = jest.fn(); - - const mockCreate = jest - .spyOn(document, 'createElement') - .mockImplementation(() => link); - - const HookWrapper: React.FC = () => { - const { saveFile } = useDataSaver('message', content); - useEffect(() => saveFile(), [saveFile]); - return null; - }; - - render(<HookWrapper />); - expect(mockCreate).toHaveBeenCalledTimes(2); - expect(link.download).toEqual('message_1616581196000.json'); - expect(link.href).toEqual(`data:text/json;charset=utf-8,${content}`); - expect(link.click).toHaveBeenCalledTimes(1); - - mockCreate.mockRestore(); - }); - it('downloads txt file', () => { + global.URL.createObjectURL = jest.fn(); const link: HTMLAnchorElement = document.createElement('a'); link.click = jest.fn(); @@ -54,8 +32,7 @@ describe('useDataSaver hook', () => { render(<HookWrapper />); expect(mockCreate).toHaveBeenCalledTimes(2); - expect(link.download).toEqual('message_1616581196000.txt'); - expect(link.href).toEqual(`data:text/json;charset=utf-8,content`); + expect(link.download).toEqual('message'); expect(link.click).toHaveBeenCalledTimes(1); mockCreate.mockRestore(); diff --git a/kafka-ui-react-app/src/lib/hooks/useDataSaver.ts b/kafka-ui-react-app/src/lib/hooks/useDataSaver.ts index 6dba2700be8..d1ff61386dc 100644 --- a/kafka-ui-react-app/src/lib/hooks/useDataSaver.ts +++ b/kafka-ui-react-app/src/lib/hooks/useDataSaver.ts @@ -1,4 +1,3 @@ -import isObject from 'lodash/isObject'; import { showSuccessAlert } from 'lib/errorHandling'; const useDataSaver = ( @@ -17,19 +16,14 @@ const useDataSaver = ( }); } }; - const saveFile = () => { - const extension = isObject(data) ? 'json' : 'txt'; - const dataStr = `data:text/json;charset=utf-8,${data}`; - const downloadAnchorNode = document.createElement('a'); - downloadAnchorNode.setAttribute('href', dataStr); - downloadAnchorNode.setAttribute( - 'download', - `${subject}_${new Date().getTime()}.${extension}` - ); - document.body.appendChild(downloadAnchorNode); - downloadAnchorNode.click(); - downloadAnchorNode.remove(); + const blob = new Blob([data as BlobPart], { type: 'text/json' }); + const elem = window.document.createElement('a'); + elem.href = window.URL.createObjectURL(blob); + elem.download = subject; + document.body.appendChild(elem); + elem.click(); + document.body.removeChild(elem); }; return { copyToClipboard, saveFile };
null
train
val
2022-11-15T13:05:06
"2022-09-29T17:44:16Z"
nitin1677
train
provectus/kafka-ui/2298_2874
provectus/kafka-ui
provectus/kafka-ui/2298
provectus/kafka-ui/2874
[ "connected" ]
0818bb0144ca4b2f3b23a0ab9e374fa8bf727901
3277186fc03c5a3a381ccf24b29924d13e7627fd
[ "@m-kay hey, unfortunately, that's currently not possible to have multiple protobuf files, but we'll implement this in the future. Thanks for the suggestion!", "related #2228" ]
[ "pls also check that protobufFiles list is not empty", "pls add `distinct()`", "looks like pipeline can be simplified to smth like\r\n```\r\nprotobufSchemas.entrySet().stream() \r\n .map(e -> Map.entry(e.getValue().toDescriptor(msgName), e.getKey()))\r\n .filter(e -> e.getKey() != null)\r\n ...\r\n```", "@iliax Unfortunately `Map#entry(K, V)` will throw an exception if the provided key (`K`) is `null`, so we'll have to check this before creating the `Map.Entry` instance.\r\n\r\nWe can remove the `filter(descriptor -> descriptor.getKey() != null)` step, though.", "ok, I see, lets rm unnecessary checks though. (`schema.getValue() != null` also looks redundant) ", "TODO drop deprecated protobuf single file property" ]
"2022-11-03T16:58:40Z"
[ "type/enhancement", "scope/backend", "status/accepted" ]
Support multiple protobuf schema files
### Discussed in https://github.com/provectus/kafka-ui/discussions/2249 <div type='discussions-op-text'> <sup>Originally posted by **m-kay** July 6, 2022</sup> According to the [docs](https://github.com/provectus/kafka-ui/blob/master/documentation/guides/Protobuf.md) it is only possible to configure one schema file. However I have multiple schema files each with a different package so I'm not even able to merge them into a single file since only one package is possible per file. Do you have any idea how I could configure kafka-ui to use multiple schema files?</div>
[ "documentation/guides/Protobuf.md", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/serdes/builtin/ProtobufFileSerde.java" ]
[ "documentation/guides/Protobuf.md", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/serdes/builtin/ProtobufFileSerde.java" ]
[ "kafka-ui-api/src/test/java/com/provectus/kafka/ui/serdes/builtin/ProtobufFileSerdeTest.java", "kafka-ui-api/src/test/resources/sensor.proto" ]
diff --git a/documentation/guides/Protobuf.md b/documentation/guides/Protobuf.md index d7c50ffb65e..93bc7133e0a 100644 --- a/documentation/guides/Protobuf.md +++ b/documentation/guides/Protobuf.md @@ -11,8 +11,12 @@ To configure Kafkaui to deserialize protobuf messages using a supplied protobuf kafka: clusters: - # Cluster configuration omitted. - # protobufFile is the path to the protobuf schema. + # protobufFile is the path to the protobuf schema. (deprecated: please use "protobufFiles") protobufFile: path/to/my.proto + # protobufFiles is the path to one or more protobuf schemas. + protobufFiles: + - path/to/my.proto + - path/to/another.proto # protobufMessageName is the default protobuf type that is used to deserilize # the message's value if the topic is not found in protobufMessageNameByTopic. protobufMessageName: my.Type1 diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/serdes/builtin/ProtobufFileSerde.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/serdes/builtin/ProtobufFileSerde.java index 6e861cfe56b..567b98584b0 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/serdes/builtin/ProtobufFileSerde.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/serdes/builtin/ProtobufFileSerde.java @@ -3,6 +3,7 @@ import com.google.common.annotations.VisibleForTesting; import com.google.protobuf.Descriptors.Descriptor; import com.google.protobuf.DynamicMessage; +import com.google.protobuf.Empty; import com.google.protobuf.util.JsonFormat; import com.provectus.kafka.ui.serde.api.DeserializeResult; import com.provectus.kafka.ui.serde.api.PropertyResolver; @@ -13,18 +14,21 @@ import io.confluent.kafka.schemaregistry.protobuf.ProtobufSchema; import io.confluent.kafka.schemaregistry.protobuf.ProtobufSchemaUtils; import java.io.ByteArrayInputStream; +import java.io.IOException; +import java.io.UncheckedIOException; import java.nio.file.Files; import java.nio.file.Path; +import java.util.Collection; import java.util.HashMap; +import java.util.List; import java.util.Map; -import java.util.Objects; import java.util.Optional; +import java.util.function.Function; import java.util.stream.Collectors; import java.util.stream.Stream; import javax.annotation.Nullable; import lombok.SneakyThrows; - public class ProtobufFileSerde implements BuiltInSerde { public static String name() { @@ -33,11 +37,11 @@ public static String name() { private static final ProtobufSchemaConverter SCHEMA_CONVERTER = new ProtobufSchemaConverter(); - private Path protobufSchemaPath; - private Map<String, Descriptor> messageDescriptorMap = new HashMap<>(); private Map<String, Descriptor> keyMessageDescriptorMap = new HashMap<>(); + private Map<Descriptor, Path> descriptorPaths = new HashMap<>(); + private Descriptor defaultMessageDescriptor; @Nullable @@ -46,8 +50,10 @@ public static String name() { @Override public boolean initOnStartup(PropertyResolver kafkaClusterProperties, PropertyResolver globalProperties) { - return kafkaClusterProperties.getProperty("protobufFile", String.class) - .isPresent(); + Optional<String> protobufFile = kafkaClusterProperties.getProperty("protobufFile", String.class); + Optional<List<String>> protobufFiles = kafkaClusterProperties.getListProperty("protobufFiles", String.class); + + return protobufFile.isPresent() || protobufFiles.map(files -> files.isEmpty() ? null : files).isPresent(); } @SneakyThrows @@ -55,55 +61,107 @@ public boolean initOnStartup(PropertyResolver kafkaClusterProperties, public void configure(PropertyResolver serdeProperties, PropertyResolver kafkaClusterProperties, PropertyResolver globalProperties) { - protobufSchemaPath = Path.of( - kafkaClusterProperties.getProperty("protobufFile", String.class) - .orElseThrow()); - ProtobufSchema protobufSchema; - try (Stream<String> lines = Files.lines(protobufSchemaPath)) { - protobufSchema = new ProtobufSchema(lines.collect(Collectors.joining("\n"))); - } + Map<Path, ProtobufSchema> protobufSchemas = joinPathProperties(kafkaClusterProperties).stream() + .map(path -> Map.entry(path, new ProtobufSchema(readFileAsString(path)))) + .collect(Collectors.toMap(Map.Entry::getKey, Map.Entry::getValue)); + + + // Load all referenced message schemas and store their source proto file with the descriptors + Map<Descriptor, Path> descriptorPaths = new HashMap<>(); + Optional<String> protobufMessageName = kafkaClusterProperties.getProperty("protobufMessageName", String.class); + protobufMessageName.ifPresent(messageName -> addProtobufSchema(descriptorPaths, protobufSchemas, messageName)); + + Optional<String> protobufMessageNameForKey = + kafkaClusterProperties.getProperty("protobufMessageNameForKey", String.class); + protobufMessageNameForKey + .ifPresent(messageName -> addProtobufSchema(descriptorPaths, protobufSchemas, messageName)); + + Optional<Map<String, String>> protobufMessageNameByTopic = + kafkaClusterProperties.getMapProperty("protobufMessageNameByTopic", String.class, String.class); + protobufMessageNameByTopic + .ifPresent(messageNamesByTopic -> addProtobufSchemas(descriptorPaths, protobufSchemas, messageNamesByTopic)); + + Optional<Map<String, String>> protobufMessageNameForKeyByTopic = + kafkaClusterProperties.getMapProperty("protobufMessageNameForKeyByTopic", String.class, String.class); + protobufMessageNameForKeyByTopic + .ifPresent(messageNamesByTopic -> addProtobufSchemas(descriptorPaths, protobufSchemas, messageNamesByTopic)); + + // Fill dictionary for descriptor lookup by full message name + Map<String, Descriptor> descriptorMap = descriptorPaths.keySet().stream() + .collect(Collectors.toMap(Descriptor::getFullName, Function.identity())); + configure( - protobufSchemaPath, - defaultMessageDescriptor = kafkaClusterProperties.getProperty("protobufMessageName", String.class) - .map(msgName -> Objects.requireNonNull(protobufSchema.toDescriptor(msgName), - "The given message type not found in protobuf definition: " + msgName)) - // this is strange logic, but we need it to support serde's backward-compatibility - .orElseGet(protobufSchema::toDescriptor), - defaultKeyMessageDescriptor = kafkaClusterProperties.getProperty("protobufMessageNameForKey", String.class) - .map(msgName -> Objects.requireNonNull(protobufSchema.toDescriptor(msgName), - "The given message type not found in protobuf definition: " + msgName)) - .orElse(null), - kafkaClusterProperties.getMapProperty("protobufMessageNameByTopic", String.class, String.class) - .map(map -> populateDescriptors(protobufSchema, map)) - .orElse(Map.of()), - kafkaClusterProperties.getMapProperty("protobufMessageNameForKeyByTopic", String.class, String.class) - .map(map -> populateDescriptors(protobufSchema, map)) - .orElse(Map.of()) + // this is strange logic, but we need it to support serde's backward-compatibility + protobufMessageName.map(descriptorMap::get).orElseGet(Empty::getDescriptor), + protobufMessageNameForKey.map(descriptorMap::get).orElse(null), + descriptorPaths, + protobufMessageNameByTopic.map(map -> populateDescriptors(descriptorMap, map)).orElse(Map.of()), + protobufMessageNameForKeyByTopic.map(map -> populateDescriptors(descriptorMap, map)).orElse(Map.of()) ); } @VisibleForTesting void configure( - Path protobufSchemaPath, Descriptor defaultMessageDescriptor, @Nullable Descriptor defaultKeyMessageDescriptor, + Map<Descriptor, Path> descriptorPaths, Map<String, Descriptor> messageDescriptorMap, Map<String, Descriptor> keyMessageDescriptorMap) { - this.protobufSchemaPath = protobufSchemaPath; this.defaultMessageDescriptor = defaultMessageDescriptor; this.defaultKeyMessageDescriptor = defaultKeyMessageDescriptor; + this.descriptorPaths = descriptorPaths; this.messageDescriptorMap = messageDescriptorMap; this.keyMessageDescriptorMap = keyMessageDescriptorMap; } - private Map<String, Descriptor> populateDescriptors(ProtobufSchema protobufSchema, + private static void addProtobufSchema(Map<Descriptor, Path> descriptorPaths, + Map<Path, ProtobufSchema> protobufSchemas, + String messageName) { + var descriptorAndPath = getDescriptorAndPath(protobufSchemas, messageName); + descriptorPaths.put(descriptorAndPath.getKey(), descriptorAndPath.getValue()); + } + + private static void addProtobufSchemas(Map<Descriptor, Path> descriptorPaths, + Map<Path, ProtobufSchema> protobufSchemas, + Map<String, String> messageNamesByTopic) { + messageNamesByTopic.values().stream() + .map(msgName -> getDescriptorAndPath(protobufSchemas, msgName)) + .forEach(entry -> descriptorPaths.put(entry.getKey(), entry.getValue())); + } + + private static List<Path> joinPathProperties(PropertyResolver propertyResolver) { + return Stream.concat( + propertyResolver.getProperty("protobufFile", String.class).map(List::of).stream(), + propertyResolver.getListProperty("protobufFiles", String.class).stream()) + .flatMap(Collection::stream) + .distinct() + .map(Path::of) + .collect(Collectors.toList()); + } + + private static Map.Entry<Descriptor, Path> getDescriptorAndPath(Map<Path, ProtobufSchema> protobufSchemas, + String msgName) { + return protobufSchemas.entrySet().stream() + .filter(schema -> schema.getValue() != null && schema.getValue().toDescriptor(msgName) != null) + .map(schema -> Map.entry(schema.getValue().toDescriptor(msgName), schema.getKey())) + .findFirst() + .orElseThrow(() -> new NullPointerException( + "The given message type not found in protobuf definition: " + msgName)); + } + + private static String readFileAsString(Path path) { + try { + return Files.readString(path); + } catch (IOException e) { + throw new UncheckedIOException(e); + } + } + + private Map<String, Descriptor> populateDescriptors(Map<String, Descriptor> descriptorMap, Map<String, String> messageNameMap) { Map<String, Descriptor> descriptors = new HashMap<>(); for (Map.Entry<String, String> entry : messageNameMap.entrySet()) { - var descriptor = Objects.requireNonNull(protobufSchema.toDescriptor(entry.getValue()), - "The given message type is not found in protobuf definition: " - + entry.getValue()); - descriptors.put(entry.getKey(), descriptor); + descriptors.put(entry.getKey(), descriptorMap.get(entry.getValue())); } return descriptors; } @@ -168,11 +226,14 @@ public DeserializeResult deserialize(RecordHeaders headers, byte[] data) { @Override public Optional<SchemaDescription> getSchema(String topic, Target type) { - return descriptorFor(topic, type) - .map(descriptor -> - new SchemaDescription( - SCHEMA_CONVERTER.convert(protobufSchemaPath.toUri(), descriptor).toJson(), - Map.of("messageName", descriptor.getFullName()) - )); + return descriptorFor(topic, type).map(this::toSchemaDescription); + } + + private SchemaDescription toSchemaDescription(Descriptor descriptor) { + Path path = descriptorPaths.get(descriptor); + return new SchemaDescription( + SCHEMA_CONVERTER.convert(path.toUri(), descriptor).toJson(), + Map.of("messageName", descriptor.getFullName()) + ); } }
diff --git a/kafka-ui-api/src/test/java/com/provectus/kafka/ui/serdes/builtin/ProtobufFileSerdeTest.java b/kafka-ui-api/src/test/java/com/provectus/kafka/ui/serdes/builtin/ProtobufFileSerdeTest.java index 062839a6c0a..887792b6335 100644 --- a/kafka-ui-api/src/test/java/com/provectus/kafka/ui/serdes/builtin/ProtobufFileSerdeTest.java +++ b/kafka-ui-api/src/test/java/com/provectus/kafka/ui/serdes/builtin/ProtobufFileSerdeTest.java @@ -1,19 +1,25 @@ package com.provectus.kafka.ui.serdes.builtin; import static org.assertj.core.api.Assertions.assertThat; +import static org.assertj.core.api.Assertions.assertThatThrownBy; +import static org.mockito.Mockito.mock; +import static org.mockito.Mockito.when; import com.fasterxml.jackson.databind.json.JsonMapper; -import com.google.protobuf.DynamicMessage; +import com.google.protobuf.Descriptors; import com.google.protobuf.util.JsonFormat; +import com.provectus.kafka.ui.serde.api.PropertyResolver; import com.provectus.kafka.ui.serde.api.Serde; import io.confluent.kafka.schemaregistry.protobuf.ProtobufSchema; import java.nio.file.Files; import java.nio.file.Path; -import java.nio.file.Paths; +import java.util.List; import java.util.Map; +import java.util.Optional; import lombok.SneakyThrows; -import org.junit.jupiter.api.BeforeAll; +import org.junit.jupiter.api.BeforeEach; import org.junit.jupiter.api.Test; +import org.springframework.util.ResourceUtils; class ProtobufFileSerdeTest { @@ -23,43 +29,93 @@ class ProtobufFileSerdeTest { private static final String sampleBookMsgJson = "{\"version\": 1, \"people\": [" + "{ \"name\": \"My Name\",\"id\": 102, \"email\": \"[email protected]\", \"phones\":[]}]}"; + private static final String sampleSensorMsgJson = "{ \"name\": \"My Sensor\", " + + "\"temperature\": 20.5, \"humidity\": 50, \"door\": \"OPEN\" }"; + // Sample message of type `test.Person` - private static byte[] personMessageBytes; + private byte[] personMessageBytes; // Sample message of type `test.AddressBook` - private static byte[] addressBookMessageBytes; - private static Path protobufSchemaPath; - private static ProtobufSchema protobufSchema; + private byte[] addressBookMessageBytes; + private byte[] sensorMessageBytes; + private Path addressBookSchemaPath; + private Path sensorSchemaPath; + + private Descriptors.Descriptor personDescriptor; + private Descriptors.Descriptor addressBookDescriptor; + private Descriptors.Descriptor sensorDescriptor; + private Map<Descriptors.Descriptor, Path> descriptorPaths; - @BeforeAll - static void setUp() throws Exception { - protobufSchemaPath = Paths.get(ProtobufFileSerdeTest.class.getClassLoader() - .getResource("address-book.proto").toURI()); - protobufSchema = new ProtobufSchema(Files.readString(protobufSchemaPath)); + @BeforeEach + void setUp() throws Exception { + addressBookSchemaPath = ResourceUtils.getFile("classpath:address-book.proto").toPath(); + sensorSchemaPath = ResourceUtils.getFile("classpath:sensor.proto").toPath(); - DynamicMessage.Builder builder = protobufSchema.newMessageBuilder("test.Person"); + ProtobufSchema addressBookSchema = new ProtobufSchema(Files.readString(addressBookSchemaPath)); + var builder = addressBookSchema.newMessageBuilder("test.Person"); JsonFormat.parser().merge(samplePersonMsgJson, builder); personMessageBytes = builder.build().toByteArray(); - builder = protobufSchema.newMessageBuilder("test.AddressBook"); + builder = addressBookSchema.newMessageBuilder("test.AddressBook"); JsonFormat.parser().merge(sampleBookMsgJson, builder); addressBookMessageBytes = builder.build().toByteArray(); + personDescriptor = addressBookSchema.toDescriptor("test.Person"); + addressBookDescriptor = addressBookSchema.toDescriptor("test.AddressBook"); + + ProtobufSchema sensorSchema = new ProtobufSchema(Files.readString(sensorSchemaPath)); + builder = sensorSchema.newMessageBuilder("iot.Sensor"); + JsonFormat.parser().merge(sampleSensorMsgJson, builder); + sensorMessageBytes = builder.build().toByteArray(); + sensorDescriptor = sensorSchema.toDescriptor("iot.Sensor"); + + descriptorPaths = Map.of( + personDescriptor, addressBookSchemaPath, + addressBookDescriptor, addressBookSchemaPath, + sensorDescriptor, sensorSchemaPath + ); } @Test void testDeserialize() { var messageNameMap = Map.of( - "persons", protobufSchema.toDescriptor("test.Person"), - "books", protobufSchema.toDescriptor("test.AddressBook") + "persons", personDescriptor, + "books", addressBookDescriptor ); var keyMessageNameMap = Map.of( - "books", protobufSchema.toDescriptor("test.AddressBook")); + "books", addressBookDescriptor); + + var serde = new ProtobufFileSerde(); + serde.configure( + null, + null, + descriptorPaths, + messageNameMap, + keyMessageNameMap + ); + + var deserializedPerson = serde.deserializer("persons", Serde.Target.VALUE) + .deserialize(null, personMessageBytes); + assertJsonEquals(samplePersonMsgJson, deserializedPerson.getResult()); + + var deserializedBook = serde.deserializer("books", Serde.Target.KEY) + .deserialize(null, addressBookMessageBytes); + assertJsonEquals(sampleBookMsgJson, deserializedBook.getResult()); + } + @Test + void testDeserializeMultipleProtobuf() { + var messageNameMap = Map.of( + "persons", personDescriptor, + "books", addressBookDescriptor, + "sensors", sensorDescriptor + ); + var keyMessageNameMap = Map.of( + "books", addressBookDescriptor); var serde = new ProtobufFileSerde(); serde.configure( - protobufSchemaPath, null, null, + descriptorPaths, messageNameMap, keyMessageNameMap ); @@ -71,15 +127,19 @@ void testDeserialize() { var deserializedBook = serde.deserializer("books", Serde.Target.KEY) .deserialize(null, addressBookMessageBytes); assertJsonEquals(sampleBookMsgJson, deserializedBook.getResult()); + + var deserializedSensor = serde.deserializer("sensors", Serde.Target.VALUE) + .deserialize(null, sensorMessageBytes); + assertJsonEquals(sampleSensorMsgJson, deserializedSensor.getResult()); } @Test void testDefaultMessageName() { var serde = new ProtobufFileSerde(); serde.configure( - protobufSchemaPath, - protobufSchema.toDescriptor("test.Person"), - protobufSchema.toDescriptor("test.AddressBook"), + personDescriptor, + addressBookDescriptor, + descriptorPaths, Map.of(), Map.of() ); @@ -93,21 +153,20 @@ void testDefaultMessageName() { assertJsonEquals(sampleBookMsgJson, deserializedBook.getResult()); } - @Test void testSerialize() { var messageNameMap = Map.of( - "persons", protobufSchema.toDescriptor("test.Person"), - "books", protobufSchema.toDescriptor("test.AddressBook") + "persons", personDescriptor, + "books", addressBookDescriptor ); var keyMessageNameMap = Map.of( - "books", protobufSchema.toDescriptor("test.AddressBook")); + "books", addressBookDescriptor); var serde = new ProtobufFileSerde(); serde.configure( - protobufSchemaPath, null, null, + descriptorPaths, messageNameMap, keyMessageNameMap ); @@ -122,13 +181,46 @@ void testSerialize() { assertThat(booksBytes).isEqualTo(addressBookMessageBytes); } + @Test + void testSerializeMultipleProtobuf() { + var messageNameMap = Map.of( + "persons", personDescriptor, + "books", addressBookDescriptor, + "sensors", sensorDescriptor + ); + var keyMessageNameMap = Map.of( + "books", addressBookDescriptor); + + var serde = new ProtobufFileSerde(); + serde.configure( + null, + null, + descriptorPaths, + messageNameMap, + keyMessageNameMap + ); + + var personBytes = serde.serializer("persons", Serde.Target.VALUE) + .serialize("{ \"name\": \"My Name\",\"id\": 101, \"email\": \"[email protected]\" }"); + assertThat(personBytes).isEqualTo(personMessageBytes); + + var booksBytes = serde.serializer("books", Serde.Target.KEY) + .serialize("{\"version\": 1, \"people\": [" + + "{ \"name\": \"My Name\",\"id\": 102, \"email\": \"[email protected]\" }]}"); + assertThat(booksBytes).isEqualTo(addressBookMessageBytes); + + var sensorBytes = serde.serializer("sensors", Serde.Target.VALUE) + .serialize("{ \"name\": \"My Sensor\", \"temperature\": 20.5, \"humidity\": 50, \"door\": \"OPEN\" }"); + assertThat(sensorBytes).isEqualTo(sensorMessageBytes); + } + @Test void testSerializeDefaults() { var serde = new ProtobufFileSerde(); serde.configure( - protobufSchemaPath, - protobufSchema.toDescriptor("test.Person"), - protobufSchema.toDescriptor("test.AddressBook"), + personDescriptor, + addressBookDescriptor, + descriptorPaths, Map.of(), Map.of() ); @@ -143,10 +235,152 @@ void testSerializeDefaults() { assertThat(booksBytes).isEqualTo(addressBookMessageBytes); } + @Test + void initOnStartupReturnsFalseIfNoProtoFilesHaveBeenProvided() { + PropertyResolver resolver = mock(PropertyResolver.class); + + var serde = new ProtobufFileSerde(); + boolean startupSuccessful = serde.initOnStartup(resolver, resolver); + assertThat(startupSuccessful).isFalse(); + } + + @Test + void initOnStartupReturnsFalseIfProtoFilesListIsEmpty() { + PropertyResolver resolver = mock(PropertyResolver.class); + when(resolver.getListProperty("protobufFiles", String.class)).thenReturn(Optional.of(List.of())); + + var serde = new ProtobufFileSerde(); + boolean startupSuccessful = serde.initOnStartup(resolver, resolver); + assertThat(startupSuccessful).isFalse(); + } + + @Test + void initOnStartupReturnsTrueIfNoProtoFileHasBeenProvided() { + PropertyResolver resolver = mock(PropertyResolver.class); + when(resolver.getProperty("protobufFile", String.class)).thenReturn(Optional.of("file.proto")); + + var serde = new ProtobufFileSerde(); + boolean startupSuccessful = serde.initOnStartup(resolver, resolver); + assertThat(startupSuccessful).isTrue(); + } + + @Test + void initOnStartupReturnsTrueIfProtoFilesHasBeenProvided() { + PropertyResolver resolver = mock(PropertyResolver.class); + when(resolver.getListProperty("protobufFiles", String.class)).thenReturn(Optional.of(List.of("file.proto"))); + + var serde = new ProtobufFileSerde(); + boolean startupSuccessful = serde.initOnStartup(resolver, resolver); + assertThat(startupSuccessful).isTrue(); + } + + @Test + void initOnStartupReturnsTrueIfProtoFileAndProtoFilesHaveBeenProvided() { + PropertyResolver resolver = mock(PropertyResolver.class); + when(resolver.getProperty("protobufFile", String.class)).thenReturn(Optional.of("file1.proto")); + when(resolver.getListProperty("protobufFiles", String.class)).thenReturn(Optional.of(List.of("file2.proto"))); + + var serde = new ProtobufFileSerde(); + boolean startupSuccessful = serde.initOnStartup(resolver, resolver); + assertThat(startupSuccessful).isTrue(); + } + + @Test + void listOfProtobufFilesIsJoined() { + PropertyResolver resolver = mock(PropertyResolver.class); + when(resolver.getProperty("protobufFile", String.class)) + .thenReturn(Optional.of(addressBookSchemaPath.toString())); + when(resolver.getListProperty("protobufFiles", String.class)) + .thenReturn(Optional.of(List.of(sensorSchemaPath.toString()))); + when(resolver.getProperty("protobufMessageName", String.class)) + .thenReturn(Optional.of("test.AddressBook")); + + Map<String, String> protobufMessageNameByTopic = Map.of( + "persons", "test.Person", + "books", "test.AddressBook", + "sensors", "iot.Sensor"); + when(resolver.getMapProperty("protobufMessageNameByTopic", String.class, String.class)) + .thenReturn(Optional.of(protobufMessageNameByTopic)); + + var serde = new ProtobufFileSerde(); + serde.configure(resolver, resolver, resolver); + + var deserializedPerson = serde.deserializer("persons", Serde.Target.VALUE) + .deserialize(null, personMessageBytes); + assertJsonEquals(samplePersonMsgJson, deserializedPerson.getResult()); + + var deserializedSensor = serde.deserializer("sensors", Serde.Target.VALUE) + .deserialize(null, sensorMessageBytes); + assertJsonEquals(sampleSensorMsgJson, deserializedSensor.getResult()); + } + + @Test + void unknownSchemaAsDefaultThrowsException() { + PropertyResolver resolver = mock(PropertyResolver.class); + when(resolver.getListProperty("protobufFiles", String.class)) + .thenReturn(Optional.of(List.of(addressBookSchemaPath.toString(), sensorSchemaPath.toString()))); + when(resolver.getProperty("protobufMessageName", String.class)) + .thenReturn(Optional.of("test.NotExistent")); + + var serde = new ProtobufFileSerde(); + assertThatThrownBy(() -> serde.configure(resolver, resolver, resolver)) + .isInstanceOf(NullPointerException.class) + .hasMessage("The given message type not found in protobuf definition: test.NotExistent"); + } + + @Test + void unknownSchemaAsDefaultForKeyThrowsException() { + PropertyResolver resolver = mock(PropertyResolver.class); + when(resolver.getListProperty("protobufFiles", String.class)) + .thenReturn(Optional.of(List.of(addressBookSchemaPath.toString(), sensorSchemaPath.toString()))); + when(resolver.getProperty("protobufMessageName", String.class)) + .thenReturn(Optional.of("test.AddressBook")); + when(resolver.getProperty("protobufMessageNameForKey", String.class)) + .thenReturn(Optional.of("test.NotExistent")); + + var serde = new ProtobufFileSerde(); + assertThatThrownBy(() -> serde.configure(resolver, resolver, resolver)) + .isInstanceOf(NullPointerException.class) + .hasMessage("The given message type not found in protobuf definition: test.NotExistent"); + } + + @Test + void unknownSchemaAsTopicSchemaThrowsException() { + PropertyResolver resolver = mock(PropertyResolver.class); + when(resolver.getListProperty("protobufFiles", String.class)) + .thenReturn(Optional.of(List.of(addressBookSchemaPath.toString(), sensorSchemaPath.toString()))); + when(resolver.getProperty("protobufMessageName", String.class)) + .thenReturn(Optional.of("test.AddressBook")); + + when(resolver.getMapProperty("protobufMessageNameByTopic", String.class, String.class)) + .thenReturn(Optional.of(Map.of("persons", "test.NotExistent"))); + + var serde = new ProtobufFileSerde(); + assertThatThrownBy(() -> serde.configure(resolver, resolver, resolver)) + .isInstanceOf(NullPointerException.class) + .hasMessage("The given message type not found in protobuf definition: test.NotExistent"); + } + + @Test + void unknownSchemaAsTopicSchemaForKeyThrowsException() { + PropertyResolver resolver = mock(PropertyResolver.class); + when(resolver.getListProperty("protobufFiles", String.class)) + .thenReturn(Optional.of(List.of(addressBookSchemaPath.toString(), sensorSchemaPath.toString()))); + when(resolver.getProperty("protobufMessageName", String.class)) + .thenReturn(Optional.of("test.AddressBook")); + + when(resolver.getMapProperty("protobufMessageNameForKeyByTopic", String.class, String.class)) + .thenReturn(Optional.of(Map.of("persons", "test.NotExistent"))); + + var serde = new ProtobufFileSerde(); + assertThatThrownBy(() -> serde.configure(resolver, resolver, resolver)) + .isInstanceOf(NullPointerException.class) + .hasMessage("The given message type not found in protobuf definition: test.NotExistent"); + } + @SneakyThrows private void assertJsonEquals(String expectedJson, String actualJson) { var mapper = new JsonMapper(); assertThat(mapper.readTree(actualJson)).isEqualTo(mapper.readTree(expectedJson)); } - -} \ No newline at end of file +} diff --git a/kafka-ui-api/src/test/resources/sensor.proto b/kafka-ui-api/src/test/resources/sensor.proto new file mode 100644 index 00000000000..33b8c387e4c --- /dev/null +++ b/kafka-ui-api/src/test/resources/sensor.proto @@ -0,0 +1,14 @@ +syntax = "proto3"; +package iot; + +message Sensor { + string name = 1; + double temperature = 2; + int32 humidity = 3; + + enum SwitchLevel { + CLOSED = 0; + OPEN = 1; + } + SwitchLevel door = 5; +}
train
val
2022-11-04T20:11:08
"2022-07-19T10:47:17Z"
Haarolean
train
provectus/kafka-ui/2703_2887
provectus/kafka-ui
provectus/kafka-ui/2703
provectus/kafka-ui/2887
[ "connected" ]
b0c897b5c804cfec7d9941438f7d2dab5b60b6e2
9f9bd36b0fbf17776bc520e1e4dc4eb85a4a0b09
[ "@Haarolean, we already return if replica is sync or not from backend. So, we can easily implement it in UI. As for design - I think the simplest way is to highlight non-synced replicates with bold red. ", "- make the whole line red in case when the leader is offline?" ]
[]
"2022-11-07T08:30:31Z"
[ "type/enhancement", "good first issue", "scope/frontend", "status/accepted" ]
Display In Sync Replicas (ISR) by Partition on Topic overview
### Which version of the app are you running? 0.4.0 for production use Tu.10.2022 06:31:55 (c961344) for test and evaluation ### Is your proposal related to a problem? It is helpful to identify which partitions for topics are not replicating properly, causing the In Sync Replicas number to be reduced in an otherwise normally functioning cluster (i.e. no brokers are down). ### Describe the solution you'd like Currently, when viewing a topic that has under replicated partitions, the metrics at the topic highlight this by show the number of URP and ISR. ![image](https://user-images.githubusercontent.com/18301427/194098806-0d6929f2-fb3c-4870-bf0f-51689a5ac988.png) On the latest build, on the Partition list for a topic, there is a Replicas column, with green showing which broker is the leader for that partition. ![image](https://user-images.githubusercontent.com/18301427/194099133-196ee221-7cad-42e2-9635-edba167b5378.png) If we could add a column that showed which replicas for each partition are in sync, it could highlight those that are out of sync, like this (taken from a different Kafka tool). ![image](https://user-images.githubusercontent.com/18301427/194099470-c2699447-f168-4cb7-8f23-2cafa596f509.png) This helps identify if all the out of sync replicas are from a single broker, which they were in my case. All URP issues were on broker 1, so we were able to focus our attention there.
[ "kafka-ui-react-app/src/components/Topics/Topic/Overview/Overview.styled.ts", "kafka-ui-react-app/src/components/Topics/Topic/Overview/Overview.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Overview/__test__/Overview.spec.tsx", "kafka-ui-react-app/src/theme/theme.ts" ]
[ "kafka-ui-react-app/src/components/Topics/Topic/Overview/Overview.styled.ts", "kafka-ui-react-app/src/components/Topics/Topic/Overview/Overview.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Overview/__test__/Overview.spec.tsx", "kafka-ui-react-app/src/theme/theme.ts" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Overview/Overview.styled.ts b/kafka-ui-react-app/src/components/Topics/Topic/Overview/Overview.styled.ts index a64db26117e..d274a1ba9d9 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Overview/Overview.styled.ts +++ b/kafka-ui-react-app/src/components/Topics/Topic/Overview/Overview.styled.ts @@ -2,9 +2,15 @@ import styled from 'styled-components'; export const Replica = styled.span.attrs({ 'aria-label': 'replica-info' })<{ leader?: boolean; + outOfSync?: boolean; }>` - color: ${({ leader, theme }) => - leader ? theme.topicMetaData.liderReplica.color : null}; + color: ${({ leader, outOfSync, theme }) => { + if (outOfSync) return theme.topicMetaData.outOfSync.color; + if (leader) return theme.topicMetaData.liderReplica.color; + return null; + }}; + + font-weight: ${({ outOfSync }) => (outOfSync ? '500' : null)}; &:after { content: ', '; diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Overview/Overview.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Overview/Overview.tsx index cada465548f..f0c2ddfd61b 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Overview/Overview.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Overview/Overview.tsx @@ -51,9 +51,10 @@ const Overview: React.FC = () => { if (replicas === undefined || replicas.length === 0) { return 0; } - return replicas?.map(({ broker, leader }: Replica) => ( + return replicas?.map(({ broker, leader, inSync }: Replica) => ( <S.Replica leader={leader} + outOfSync={!inSync} key={broker} title={leader ? 'Leader' : ''} > diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Overview/__test__/Overview.spec.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Overview/__test__/Overview.spec.tsx index 3f10635df52..0857de9a6f5 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Overview/__test__/Overview.spec.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Overview/__test__/Overview.spec.tsx @@ -69,6 +69,19 @@ describe('Overview', () => { ); }); + describe('when replicas out of sync', () => { + it('should be the appropriate color', () => { + render(<Replica outOfSync />); + const element = screen.getByLabelText('replica-info'); + expect(element).toBeInTheDocument(); + expect(element).toHaveStyleRule( + 'color', + theme.topicMetaData.outOfSync.color + ); + expect(element).toHaveStyleRule('font-weight', '500'); + }); + }); + describe('when it has internal flag', () => { it('renders the Action button for Topic', () => { renderComponent({ diff --git a/kafka-ui-react-app/src/theme/theme.ts b/kafka-ui-react-app/src/theme/theme.ts index 33c8b2959c8..8684f043774 100644 --- a/kafka-ui-react-app/src/theme/theme.ts +++ b/kafka-ui-react-app/src/theme/theme.ts @@ -516,6 +516,9 @@ const theme = { liderReplica: { color: Colors.green[60], }, + outOfSync: { + color: Colors.red[50], + }, }, dangerZone: { borderColor: Colors.neutral[10],
null
train
val
2022-12-09T16:09:13
"2022-10-05T15:27:56Z"
DaveSlinn
train
provectus/kafka-ui/2883_2888
provectus/kafka-ui
provectus/kafka-ui/2883
provectus/kafka-ui/2888
[ "connected" ]
0818bb0144ca4b2f3b23a0ab9e374fa8bf727901
70cdeea819e812ccf38bb1daa1dd27ba6296d28a
[]
[ "isCreateTopicButtonEnabled()", "Changed." ]
"2022-11-07T09:07:51Z"
[ "scope/QA", "scope/AQA" ]
[e2e] Checking Topic creation possibility in case of empty Topic Name
Autotest implementation for: https://app.qase.io/project/KAFKAUI?view=1&suite=4&case=4 Description: Checking Creation of a new Topic with letting required field empty Pre-conditions: Login to Kafka-ui application Open the 'Local' section Select the 'Topics' Redirect to 'Create new Topic' page with click on '+ Add a Topic' button from 'All Topics' page Steps: 1 Let 'Topic Name *' field empty 2 Fill 'Topic name* field "testTopic1" Expected result: 1 "Create topic" button should stay disabled 2 "Create topic" button should become enabled
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicCreateEditForm.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicCreateEditForm.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicCreateEditForm.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicCreateEditForm.java index 539520445ed..9a97e6dba32 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicCreateEditForm.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicCreateEditForm.java @@ -8,6 +8,7 @@ import static com.codeborne.selenide.Selenide.*; import static com.provectus.kafka.ui.utilities.WebUtils.clickByJavaScript; +import static com.provectus.kafka.ui.utilities.WebUtils.isEnabled; import static org.assertj.core.api.Assertions.assertThat; public class TopicCreateEditForm { @@ -269,6 +270,10 @@ public String getVisibleText() { } } + public boolean isCreateTopicButtonEnabled(){ + return isEnabled(createTopicBtn); + } + private TopicCreateEditForm selectFromDropDownByOptionValue(String dropDownElementName, String optionValue) { KafkaUISelectElement select = new KafkaUISelectElement(dropDownElementName); diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java index ad93bfbc772..e35cad9769a 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java @@ -46,4 +46,15 @@ public static boolean isVisible(SelenideElement element) { return isVisible; } -} \ No newline at end of file + public static boolean isEnabled(SelenideElement element) { + log.debug("\nisEnabled: {}", element.getSearchCriteria()); + boolean isEnabled = false; + try { + element.shouldBe(Condition.enabled); + isEnabled = true; + } catch (Throwable e) { + log.debug("{} is not enabled", element.getSearchCriteria()); + } + return isEnabled; + } +}
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java index 97f70c194e3..07bdf3b0b52 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java @@ -1,5 +1,14 @@ package com.provectus.kafka.ui.tests; +import static com.provectus.kafka.ui.pages.NaviSideBar.SideMenuOption.TOPICS; +import static com.provectus.kafka.ui.pages.topic.TopicCreateEditForm.CleanupPolicyValue.COMPACT; +import static com.provectus.kafka.ui.pages.topic.TopicCreateEditForm.CleanupPolicyValue.DELETE; +import static com.provectus.kafka.ui.pages.topic.TopicCreateEditForm.MaxSizeOnDisk.SIZE_20_GB; +import static com.provectus.kafka.ui.settings.Source.CLUSTER_NAME; +import static com.provectus.kafka.ui.utilities.FileUtils.fileToString; +import static org.apache.commons.lang.RandomStringUtils.randomAlphabetic; +import static org.assertj.core.api.Assertions.assertThat; + import com.provectus.kafka.ui.base.BaseTest; import com.provectus.kafka.ui.models.Topic; import com.provectus.kafka.ui.pages.topic.TopicDetails; @@ -8,20 +17,16 @@ import com.provectus.kafka.ui.utilities.qaseIoUtils.enums.Status; import io.qameta.allure.Issue; import io.qase.api.annotation.CaseId; -import org.assertj.core.api.SoftAssertions; -import org.junit.jupiter.api.*; - import java.util.ArrayList; import java.util.List; - -import static com.provectus.kafka.ui.pages.NaviSideBar.SideMenuOption.TOPICS; -import static com.provectus.kafka.ui.pages.topic.TopicCreateEditForm.CleanupPolicyValue.COMPACT; -import static com.provectus.kafka.ui.pages.topic.TopicCreateEditForm.CleanupPolicyValue.DELETE; -import static com.provectus.kafka.ui.pages.topic.TopicCreateEditForm.MaxSizeOnDisk.SIZE_20_GB; -import static com.provectus.kafka.ui.settings.Source.CLUSTER_NAME; -import static com.provectus.kafka.ui.utilities.FileUtils.fileToString; -import static org.apache.commons.lang.RandomStringUtils.randomAlphabetic; -import static org.assertj.core.api.Assertions.assertThat; +import org.assertj.core.api.SoftAssertions; +import org.junit.jupiter.api.AfterAll; +import org.junit.jupiter.api.Assertions; +import org.junit.jupiter.api.BeforeAll; +import org.junit.jupiter.api.Disabled; +import org.junit.jupiter.api.DisplayName; +import org.junit.jupiter.api.Test; +import org.junit.jupiter.api.TestInstance; @TestInstance(TestInstance.Lifecycle.PER_CLASS) public class TopicTests extends BaseTest { @@ -239,6 +244,26 @@ void redirectToConsumerFromTopic() { .withFailMessage("isTopicInConsumersDetailsVisible").isTrue(); } + @DisplayName("Checking Topic creation possibility in case of empty Topic Name") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(4) + @Test + void checkTopicCreatePossibility() { + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .clickAddTopicBtn(); + topicCreateEditForm + .waitUntilScreenReady() + .setTopicName(""); + assertThat(topicCreateEditForm.isCreateTopicButtonEnabled()).as("isCreateTopicButtonEnabled()").isFalse(); + topicCreateEditForm + .setTopicName("testTopic1"); + assertThat(topicCreateEditForm.isCreateTopicButtonEnabled()).as("isCreateTopicButtonEnabled()").isTrue(); + } + @AfterAll public void afterAll() { TOPIC_LIST.forEach(topic -> apiHelper.deleteTopic(CLUSTER_NAME, topic.getName()));
train
val
2022-11-04T20:11:08
"2022-11-04T12:24:06Z"
ArthurNiedial
train
provectus/kafka-ui/2858_2895
provectus/kafka-ui
provectus/kafka-ui/2858
provectus/kafka-ui/2895
[ "keyword_pr_to_issue" ]
20ecc74dd9aba38156e5cc88847e3a41fd663feb
dbdf14312240e0b36f378b3c274202ffe6cb818e
[ "as a workaround we could display a short month name instead and keep the field the same width", "@Haarolean \r\nhttps://github.com/provectus/kafka-ui/blob/70cdeea819e812ccf38bb1daa1dd27ba6296d28a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx#L422-L430\r\n\r\n```jsx\r\ndateFormat=\"MMM D, YYYY HH:mm\" //Nov 7, 2022 10:45\r\n```\r\nshould be okay?", "> @Haarolean\r\n> \r\n> https://github.com/provectus/kafka-ui/blob/70cdeea819e812ccf38bb1daa1dd27ba6296d28a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx#L422-L430\r\n> \r\n> ```js\r\n> dateFormat=\"MMM D, YYYY HH:mm\" //Nov 7, 2022 10:45\r\n> ```\r\n> \r\n> should be okay?\r\n\r\nyeah, thanks" ]
[ "```suggestion\r\n dateFormat=\"MMM d, yyyy HH:mm\"\r\n```" ]
"2022-11-07T12:04:16Z"
[ "good first issue", "scope/frontend", "status/accepted", "type/chore" ]
[UI] Timestamp value overlaps with dropdown field icon within Topic/Messages
**Describe the bug** The selected value of Timestamp overlaps with dropdown field icon within Topic/Messages **Set up** https://www.kafka-ui.provectus.io/ **Steps to Reproduce** Steps to reproduce the behavior: 1. Navigate to Topic/Messages tab 2. Select Timestamp for Seek Type field 3. Choose the date for Timestamp filed **Expected behavior** Dropdown filed icon should not overlap with selected value **Screenshots** <img width="1728" alt="timestamp value" src="https://user-images.githubusercontent.com/104780608/199179175-1554c741-48cc-4623-a31a-5693b41ba8ac.png">
[ "kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx" ]
[ "kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx index 1b99650f715..9f54e56d5a4 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx @@ -428,7 +428,7 @@ const Filters: React.FC<FiltersProps> = ({ onChange={(date: Date | null) => setTimestamp(date)} showTimeInput timeInputLabel="Time:" - dateFormat="MMMM d, yyyy HH:mm" + dateFormat="MMM d, yyyy HH:mm" placeholderText="Select timestamp" disabled={isTailing} />
null
train
val
2022-11-28T09:54:07
"2022-11-01T07:15:45Z"
armenuikafka
train
provectus/kafka-ui/2439_2896
provectus/kafka-ui
provectus/kafka-ui/2439
provectus/kafka-ui/2896
[ "connected" ]
8011f60ccfea888951935b9312a5f79740c09818
c86a429a172d68bb7b6109e82e762215cdf78fba
[ "+ documentation #2432" ]
[ "can we add some logging here with warn level please? so the folks who run this locally won't get frustrated" ]
"2022-11-07T13:55:53Z"
[ "scope/QA", "status/accepted", "type/chore", "scope/AQA" ]
Make e2e tests possible to run without qaseio token
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/qaseIoUtils/QaseExtension.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/qaseIoUtils/QaseExtension.java" ]
[]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/qaseIoUtils/QaseExtension.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/qaseIoUtils/QaseExtension.java index 52c3ff9b908..474acd10999 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/qaseIoUtils/QaseExtension.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/qaseIoUtils/QaseExtension.java @@ -35,21 +35,22 @@ public class QaseExtension implements TestExecutionListener { private final ResultsApi resultsApi = new ResultsApi(apiClient); private final Map<TestIdentifier, Long> testStartTimes = new ConcurrentHashMap<>(); private static final String QASE_PROJECT = "KAFKAUI"; - private static final String QASE_ENABLE = "true"; static { String qaseApiToken = System.getProperty("QASEIO_API_TOKEN"); - if (qaseApiToken == null || StringUtils.isEmpty(qaseApiToken)) { - throw new RuntimeException("QaseIO API token should be present"); - } - if ("true".equalsIgnoreCase(System.getProperty("QASEIO_CREATE_TESTRUN"))) { - System.setProperty("QASE_RUN_NAME", "Automation run " + - new SimpleDateFormat("dd/MM/yyyy HH:mm:ss").format(new Date())); + if (StringUtils.isEmpty(qaseApiToken)) { + log.warn("QASEIO_API_TOKEN system property is not set. Support for Qase will be disabled."); + System.setProperty("QASE_ENABLE", "false"); + } else { + System.setProperty("QASE_ENABLE", "true"); + System.setProperty("QASE_PROJECT_CODE", QASE_PROJECT); + System.setProperty("QASE_API_TOKEN", qaseApiToken); + System.setProperty("QASE_USE_BULK", "false"); + if ("true".equalsIgnoreCase(System.getProperty("QASEIO_CREATE_TESTRUN"))) { + System.setProperty("QASE_RUN_NAME", "Automation run " + + new SimpleDateFormat("dd/MM/yyyy HH:mm:ss").format(new Date())); + } } - System.setProperty("QASE_ENABLE", QASE_ENABLE); - System.setProperty("QASE_PROJECT_CODE", QASE_PROJECT); - System.setProperty("QASE_API_TOKEN", qaseApiToken); - System.setProperty("QASE_USE_BULK", "false"); } @Override
null
val
val
2022-11-07T16:54:11
"2022-08-12T11:14:21Z"
Haarolean
train
provectus/kafka-ui/2870_2900
provectus/kafka-ui
provectus/kafka-ui/2870
provectus/kafka-ui/2900
[ "connected" ]
8011f60ccfea888951935b9312a5f79740c09818
212479d98f79aac5f72a386125bde3b02e549463
[ "Don't forget v2" ]
[ "i think here the query have to be more specific cause it it finding many and that is why it is failing the tests" ]
"2022-11-08T06:12:05Z"
[ "type/bug", "scope/frontend", "status/accepted", "status/confirmed" ]
"Select filter" button is active without selecting any saved filter within Topic/Messages/Saved filters
**Describe the bug** "Select filter" is active in case of not selected filter within Saved filters in Topic/Messages **Set up** https://www.kafka-ui.provectus.io/ **Steps to Reproduce** Steps to reproduce the behavior: 1. Navigate to Topic/Messages 2. Press "+Add Filters" 3. Click on "Saved Filters" 4. Press active "Select filter" **Expected behavior** "Select filter" should not be functional until selecting any saved filter **Screenshots** https://user-images.githubusercontent.com/104780608/199660167-7aed650d-2527-43f3-8020-d01826e28c44.mov
[ "kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/SavedFilters.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/__tests__/SavedFilters.spec.tsx" ]
[ "kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/SavedFilters.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/__tests__/SavedFilters.spec.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/SavedFilters.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/SavedFilters.tsx index 6a757d34383..da5033e8cac 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/SavedFilters.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/SavedFilters.tsx @@ -51,6 +51,7 @@ const SavedFilters: FC<Props> = ({ </>, () => { deleteFilter(index); + setSelectedFilter(-1); } ); }; @@ -95,6 +96,7 @@ const SavedFilters: FC<Props> = ({ buttonType="primary" type="button" onClick={activateFilter} + disabled={selectedFilter === -1} > Select filter </Button> diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/__tests__/SavedFilters.spec.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/__tests__/SavedFilters.spec.tsx index dff30487862..bc0f65befd6 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/__tests__/SavedFilters.spec.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/__tests__/SavedFilters.spec.tsx @@ -51,6 +51,9 @@ describe('SavedFilter Component', () => { it('should check the rendering of the empty filter', () => { expect(screen.getByText(/no saved filter/i)).toBeInTheDocument(); expect(screen.queryByRole('savedFilter')).not.toBeInTheDocument(); + + const selectFilterButton = screen.getByText(/Select filter/i); + expect(selectFilterButton).toBeDisabled(); }); }); @@ -117,7 +120,6 @@ describe('SavedFilter Component', () => { }); it('Open Confirmation for the deletion modal', async () => { - setUpComponent({ deleteFilter: deleteMock }); const savedFilters = getSavedFilters(); const deleteIcons = screen.getAllByText('mock-DeleteIcon'); await userEvent.hover(savedFilters[0]); @@ -130,7 +132,6 @@ describe('SavedFilter Component', () => { }); it('Close Confirmations deletion modal with button', async () => { - setUpComponent({ deleteFilter: deleteMock }); const savedFilters = getSavedFilters(); const deleteIcons = screen.getAllByText('mock-DeleteIcon'); @@ -147,7 +148,6 @@ describe('SavedFilter Component', () => { }); it('Delete the saved filter', async () => { - setUpComponent({ deleteFilter: deleteMock }); const savedFilters = getSavedFilters(); const deleteIcons = screen.getAllByText('mock-DeleteIcon'); @@ -158,6 +158,9 @@ describe('SavedFilter Component', () => { await userEvent.click(screen.getByRole('button', { name: 'Confirm' })); expect(deleteMock).toHaveBeenCalledTimes(1); expect(screen.queryByRole('dialog')).not.toBeInTheDocument(); + + const selectFilterButton = screen.getByText(/Select filter/i); + expect(selectFilterButton).toBeDisabled(); }); }); });
null
train
val
2022-11-07T16:54:11
"2022-11-03T06:57:41Z"
armenuikafka
train
provectus/kafka-ui/2859_2901
provectus/kafka-ui
provectus/kafka-ui/2859
provectus/kafka-ui/2901
[ "connected" ]
9f5d9126dacb1ef8c8299d14c1a8e2efe408154d
af1065fca875f47487c64bc9b7f3598b8144b8cb
[]
[]
"2022-11-08T06:33:52Z"
[ "scope/QA", "scope/AQA" ]
[e2e] TopicTests.messageCopyFromTopicProfile : Message copy from topic profile
Autotest implementation for: https://app.qase.io/project/KAFKAUI?view=1&suite=5&previewMode=side&case=21 Description: Checking possibility to copy message within Topic/Messages Preconditions: Login to Kafka-ui application Open the 'Local' section Select the 'Topics' Open the Topic profile which has messages Turn to Messages tab Steps: 1 Hover on Message ER: Menu icon should display 2 Click on menu icon for a Message 3 Select 'Copy to clipboard' Expected result: 1 Success message should appear 2 The message content should be copied and possible to paste (For Manual testting only)
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java index c9469fab735..61a04115fb4 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java @@ -1,159 +1,230 @@ package com.provectus.kafka.ui.pages.topic; import static com.codeborne.selenide.Selenide.$; +import static com.codeborne.selenide.Selenide.$$x; import static com.codeborne.selenide.Selenide.$x; +import com.codeborne.selenide.CollectionCondition; import com.codeborne.selenide.Condition; +import com.codeborne.selenide.ElementsCollection; import com.codeborne.selenide.SelenideElement; import com.provectus.kafka.ui.pages.BasePage; import com.provectus.kafka.ui.utilities.WaitUtils; import io.qameta.allure.Step; -import java.util.Arrays; +import java.util.ArrayList; +import java.util.List; import lombok.experimental.ExtensionMethod; import org.openqa.selenium.By; @ExtensionMethod({WaitUtils.class}) public class TopicDetails extends BasePage { - protected SelenideElement clearMessagesBtn = $x(("//div[contains(text(), 'Clear messages')]")); - protected SelenideElement overviewTab = $x("//a[contains(text(),'Overview')]"); - protected SelenideElement messagesTab = $x("//a[contains(text(),'Messages')]"); - protected SelenideElement editSettingsMenu = $x("//li[@role][contains(text(),'Edit settings')]"); - protected SelenideElement removeTopicBtn = $x("//ul[@role='menu']//div[contains(text(),'Remove Topic')]"); - protected SelenideElement confirmBtn = $x("//div[@role='dialog']//button[contains(text(),'Confirm')]"); - protected SelenideElement produceMessageBtn = $x("//div//button[text()='Produce Message']"); - protected SelenideElement contentMessageTab = $x("//html//div[@id='root']/div/main//table//p"); - protected SelenideElement cleanUpPolicyField = $x("//div[contains(text(),'Clean Up Policy')]/../span/*"); - protected SelenideElement partitionsField = $x("//div[contains(text(),'Partitions')]/../span"); - protected String consumerIdLocator = "//a[@title='%s']"; - protected String topicHeaderLocator = "//h1[contains(text(),'%s')]"; - - @Step - public TopicDetails waitUntilScreenReady() { - waitUntilSpinnerDisappear(); - Arrays.asList(overviewTab,messagesTab).forEach(element -> element.shouldBe(Condition.visible)); - return this; + protected SelenideElement clearMessagesBtn = $x(("//div[contains(text(), 'Clear messages')]")); + protected SelenideElement messageAmountCell = $x("//tbody/tr/td[5]"); + protected SelenideElement overviewTab = $x("//a[contains(text(),'Overview')]"); + protected SelenideElement messagesTab = $x("//a[contains(text(),'Messages')]"); + protected SelenideElement editSettingsMenu = $x("//li[@role][contains(text(),'Edit settings')]"); + protected SelenideElement removeTopicBtn = $x("//ul[@role='menu']//div[contains(text(),'Remove Topic')]"); + protected SelenideElement confirmBtn = $x("//div[@role='dialog']//button[contains(text(),'Confirm')]"); + protected SelenideElement produceMessageBtn = $x("//div//button[text()='Produce Message']"); + protected SelenideElement contentMessageTab = $x("//html//div[@id='root']/div/main//table//p"); + protected SelenideElement cleanUpPolicyField = $x("//div[contains(text(),'Clean Up Policy')]/../span/*"); + protected SelenideElement partitionsField = $x("//div[contains(text(),'Partitions')]/../span"); + protected ElementsCollection messageGridItems = $$x("//tbody//tr"); + protected String consumerIdLocator = "//a[@title='%s']"; + protected String topicHeaderLocator = "//h1[contains(text(),'%s')]"; + + @Step + public TopicDetails waitUntilScreenReady() { + waitUntilSpinnerDisappear(); + dotMenuBtn.shouldBe(Condition.visible); + return this; + } + + @Step + public TopicDetails openDetailsTab(TopicMenu menu) { + $(By.linkText(menu.toString())).shouldBe(Condition.visible).click(); + waitUntilSpinnerDisappear(); + return this; + } + + @Step + public TopicDetails openDotMenu() { + clickByJavaScript(dotMenuBtn); + return this; + } + + @Step + public boolean isAlertWithMessageVisible(AlertHeader header, String message) { + return isAlertVisible(header, message); + } + + @Step + public TopicDetails clickEditSettingsMenu() { + editSettingsMenu.shouldBe(Condition.visible).click(); + return this; + } + + @Step + public TopicDetails clickClearMessagesMenu() { + clearMessagesBtn.shouldBe(Condition.visible).click(); + return this; + } + + @Step + public String getCleanUpPolicy() { + return cleanUpPolicyField.getText(); + } + + @Step + public String getPartitions() { + return partitionsField.getText(); + } + + @Step + public boolean isTopicHeaderVisible(String topicName) { + return isVisible($x(String.format(topicHeaderLocator, topicName))); + } + + @Step + public TopicDetails clickDeleteTopicMenu() { + removeTopicBtn.shouldBe(Condition.visible).click(); + return this; + } + + @Step + public TopicDetails clickConfirmDeleteBtn() { + confirmBtn.shouldBe(Condition.enabled).click(); + confirmBtn.shouldBe(Condition.disappear); + return this; + } + + @Step + public TopicDetails clickProduceMessageBtn() { + clickByJavaScript(produceMessageBtn); + return this; + } + + @Step + public TopicDetails openConsumerGroup(String consumerId) { + $x(String.format(consumerIdLocator, consumerId)).click(); + return this; + } + + @Step + public boolean isKeyMessageVisible(String keyMessage) { + return keyMessage.equals($("td[title]").getText()); + } + + @Step + public boolean isContentMessageVisible(String contentMessage) { + return contentMessage.matches(contentMessageTab.getText().trim()); + } + + @Step + public int getMessageCountAmount() { + return Integer.parseInt(messageAmountCell.getText().trim()); + } + + private List<TopicDetails.MessageGridItem> initItems() { + List<TopicDetails.MessageGridItem> gridItemList = new ArrayList<>(); + messageGridItems.shouldHave(CollectionCondition.sizeGreaterThan(0)) + .forEach(item -> gridItemList.add(new TopicDetails.MessageGridItem(item))); + return gridItemList; + } + + @Step + public TopicDetails.MessageGridItem getMessage(int offset) { + return initItems().stream() + .filter(e -> e.getOffset() == offset) + .findFirst().orElse(null); + } + + @Step + public TopicDetails.MessageGridItem getRandomMessage() { + return getMessage(initItems().size() - 1); + } + + public static class MessageGridItem extends BasePage { + + private final SelenideElement element; + + private MessageGridItem(SelenideElement element) { + this.element = element; } @Step - public TopicDetails openDetailsTab(TopicMenu menu) { - $(By.linkText(menu.getValue())).shouldBe(Condition.visible).click(); - return this; + public MessageGridItem clickExpand() { + clickByJavaScript(element.$x("./td[1]/span")); + return this; } - @Step - public TopicDetails openDotMenu() { - clickByJavaScript(dotMenuBtn); - return this; + private SelenideElement getOffsetElm() { + return element.$x("./td[2]"); } @Step - public TopicDetails clickEditSettingsMenu() { - editSettingsMenu.shouldBe(Condition.visible).click(); - return this; + public int getOffset() { + return Integer.parseInt(getOffsetElm().getText().trim()); } @Step - public TopicDetails clickClearMessagesMenu() { - clearMessagesBtn.shouldBe(Condition.visible).click(); - return this; + public int getPartition() { + return Integer.parseInt(element.$x("./td[3]").getText().trim()); } @Step - public String getCleanUpPolicy(){ - return cleanUpPolicyField.getText(); + public String getTimestamp() { + return element.$x("./td[4]/div").getText().trim(); } @Step - public String getPartitions(){ - return partitionsField.getText(); + public String getKey() { + return element.$x("./td[5]").getText().trim(); } @Step - public boolean isTopicHeaderVisible(String topicName) { - return isVisible($x(String.format(topicHeaderLocator,topicName))); + public String getValue() { + return element.$x("./td[6]/span/p").getText().trim(); } @Step - public TopicDetails clickDeleteTopicMenu() { - removeTopicBtn.shouldBe(Condition.visible).click(); - return this; + public MessageGridItem openDotMenu() { + getOffsetElm().hover(); + element.$x("./td[7]/div/button[@aria-label='Dropdown Toggle']") + .shouldBe(Condition.visible).click(); + return this; } @Step - public TopicDetails clickConfirmDeleteBtn() { - confirmBtn.shouldBe(Condition.enabled).click(); - confirmBtn.shouldBe(Condition.disappear); - return this; + public MessageGridItem clickCopyToClipBoard() { + clickByJavaScript(element.$x("./td[7]//li[text() = 'Copy to clipboard']") + .shouldBe(Condition.visible)); + return this; } @Step - public TopicDetails clickProduceMessageBtn() { - clickByJavaScript(produceMessageBtn); - return this; + public MessageGridItem clickSaveAsFile() { + clickByJavaScript(element.$x("./td[7]//li[text() = 'Save as a file']") + .shouldBe(Condition.visible)); + return this; } + } - @Step - public TopicDetails openConsumerGroup(String consumerId) { - $x(String.format(consumerIdLocator, consumerId)).click(); - return this; - } + public enum TopicMenu { + OVERVIEW("Overview"), + MESSAGES("Messages"), + CONSUMERS("Consumers"), + SETTINGS("Settings"); - @Step - public boolean isKeyMessageVisible(String keyMessage) { - return keyMessage.equals($("td[title]").getText()); - } + private final String value; - @Step - public boolean isContentMessageVisible(String contentMessage) { - return contentMessage.matches(contentMessageTab.getText().trim()); - } - - @Step - public String MessageCountAmount() { - return $(By.xpath("//table[@class=\"sc-hiSbEG cvnuic\"]/tbody/tr/td[5]")).getText(); + TopicMenu(String value) { + this.value = value; } - private enum DotMenuHeaderItems { - EDIT_SETTINGS("Edit settings"), - CLEAR_MESSAGES("Clear messages"), - REMOVE_TOPIC("Remove topic"); - - private final String value; - - DotMenuHeaderItems(String value) { - this.value = value; - } - - public String getValue() { - return value; - } - - @Override - public String toString() { - return "DotMenuHeaderItems{" + "value='" + value + '\'' + '}'; - } - } - - public enum TopicMenu { - OVERVIEW("Overview"), - MESSAGES("Messages"), - CONSUMERS("Consumers"), - SETTINGS("Settings"); - - private final String value; - - TopicMenu(String value) { - this.value = value; - } - - public String getValue() { - return value; - } - - @Override - public String toString() { - return "TopicMenu{" + "value='" + value + '\'' + '}'; - } + public String toString() { + return value; } + } }
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java index aaa7a6fe8b6..dc7a3b130a8 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java @@ -1,5 +1,6 @@ package com.provectus.kafka.ui.tests; +import static com.provectus.kafka.ui.pages.BasePage.AlertHeader.SUCCESS; import static com.provectus.kafka.ui.pages.NaviSideBar.SideMenuOption.TOPICS; import static com.provectus.kafka.ui.pages.topic.enums.CleanupPolicyValue.COMPACT; import static com.provectus.kafka.ui.pages.topic.enums.CleanupPolicyValue.DELETE; @@ -12,6 +13,7 @@ import com.provectus.kafka.ui.base.BaseTest; import com.provectus.kafka.ui.models.Topic; +import com.provectus.kafka.ui.pages.BasePage; import com.provectus.kafka.ui.pages.topic.TopicDetails; import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.AutomationStatus; import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.Suite; @@ -98,7 +100,8 @@ public void createTopic() { TOPIC_LIST.add(TOPIC_TO_CREATE); } - @Disabled("https://github.com/provectus/kafka-ui/issues/2625") + @Disabled() + @Issue("https://github.com/provectus/kafka-ui/issues/2625") @DisplayName("should update a topic") @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) @AutomationStatus(status = Status.AUTOMATED) @@ -193,37 +196,37 @@ void produceMessage() { softly.assertAll(); } - @Issue("Uncomment last assertion after bug https://github.com/provectus/kafka-ui/issues/2778 fix") + @Disabled + @Issue("https://github.com/provectus/kafka-ui/issues/2778") @DisplayName("clear message") @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) @AutomationStatus(status = Status.AUTOMATED) @CaseId(19) @Test void clearMessage() { - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady() - .openTopic(TOPIC_FOR_MESSAGES.getName()); - topicDetails - .waitUntilScreenReady() - .openDetailsTab(TopicDetails.TopicMenu.OVERVIEW) - .clickProduceMessageBtn(); - produceMessagePanel - .waitUntilScreenReady() - .setContentFiled(TOPIC_FOR_MESSAGES.getMessageContent()) - .setKeyField(TOPIC_FOR_MESSAGES.getMessageKey()) - .submitProduceMessage(); - topicDetails - .waitUntilScreenReady(); - String messageAmount = topicDetails.MessageCountAmount(); - assertThat(messageAmount) - .withFailMessage("message amount not equals").isEqualTo(topicDetails.MessageCountAmount()); - topicDetails - .openDotMenu() - .clickClearMessagesMenu(); -// assertThat(Integer.toString(Integer.valueOf(messageAmount)-1)) -// .withFailMessage("message amount not decrease by one").isEqualTo(topicDetails.MessageCountAmount()); + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .openTopic(TOPIC_FOR_MESSAGES.getName()); + topicDetails + .waitUntilScreenReady() + .openDetailsTab(TopicDetails.TopicMenu.OVERVIEW) + .clickProduceMessageBtn(); + int messageAmount = topicDetails.getMessageCountAmount(); + produceMessagePanel + .waitUntilScreenReady() + .setContentFiled(TOPIC_FOR_MESSAGES.getMessageContent()) + .setKeyField(TOPIC_FOR_MESSAGES.getMessageKey()) + .submitProduceMessage(); + topicDetails + .waitUntilScreenReady(); + Assertions.assertEquals(messageAmount + 1, topicDetails.getMessageCountAmount(), "getMessageCountAmount()"); + topicDetails + .openDotMenu() + .clickClearMessagesMenu() + .waitUntilScreenReady(); + Assertions.assertEquals(0, topicDetails.getMessageCountAmount(), "getMessageCountAmount()"); } @DisplayName("Redirect to consumer from topic profile") @@ -295,6 +298,30 @@ void checkCustomParametersWithinCreateNewTopic() { .as("isValidationMessageCustomParameterValueVisible()").isTrue(); } + @Disabled + @Issue("https://github.com/provectus/kafka-ui/issues/2819") + @DisplayName("Message copy from topic profile") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(21) + @Test + void copyMessageFromTopicProfile() { + String topicName = "_schemas"; + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .openTopic(topicName); + topicDetails + .waitUntilScreenReady() + .openDetailsTab(TopicDetails.TopicMenu.MESSAGES) + .getRandomMessage() + .openDotMenu() + .clickCopyToClipBoard(); + Assertions.assertTrue(topicDetails.isAlertWithMessageVisible(SUCCESS, "Copied successfully!"), + "isAlertWithMessageVisible()"); + } + @AfterAll public void afterAll() { TOPIC_LIST.forEach(topic -> apiHelper.deleteTopic(CLUSTER_NAME, topic.getName()));
val
val
2022-11-09T21:43:18
"2022-11-01T12:32:58Z"
Redbutton18
train
provectus/kafka-ui/2738_2903
provectus/kafka-ui
provectus/kafka-ui/2738
provectus/kafka-ui/2903
[ "connected" ]
fdde380072f8a269607f6ac0ef683813f31a0141
e07ce28b2c021e4e572f639be7ff6b7c10bf9cb2
[]
[ "please check the Linter warnings", "i think there with the new version of user event , you can automatically await the userEvent ", "fixed", "fixed", "why are we deleting the empty table case ?", "fixed" ]
"2022-11-08T11:31:30Z"
[ "type/enhancement", "good first issue", "scope/frontend", "status/accepted" ]
UI now showing broker list when DISABLELOGDIRSCOLLECTION=true
Currently, when log dir collection is disabled, UI not showing brokers list - I think it can be useful to see broker list anyway. Lets render list, but hide/(fill with N/A) columns with no values : segmentSize, segmentsCount. CC @Haarolean sample: <img width="1528" alt="Screenshot 2022-10-13 at 10 35 40" src="https://user-images.githubusercontent.com/702205/195522479-de35a587-14f6-491f-a396-7da93e9be629.png"> How to reproduce locally: run kafka-ui.yaml docker compose with `KAFKA_CLUSTERS_0_DISABLELOGDIRSCOLLECTION: 'true'` property added.
[ "kafka-ui-react-app/src/components/Brokers/BrokersList/BrokersList.tsx", "kafka-ui-react-app/src/components/Brokers/BrokersList/__test__/BrokersList.spec.tsx" ]
[ "kafka-ui-react-app/src/components/Brokers/BrokersList/BrokersList.tsx", "kafka-ui-react-app/src/components/Brokers/BrokersList/__test__/BrokersList.spec.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Brokers/BrokersList/BrokersList.tsx b/kafka-ui-react-app/src/components/Brokers/BrokersList/BrokersList.tsx index 3fce89fbae9..d74c1bcb700 100644 --- a/kafka-ui-react-app/src/components/Brokers/BrokersList/BrokersList.tsx +++ b/kafka-ui-react-app/src/components/Brokers/BrokersList/BrokersList.tsx @@ -10,6 +10,8 @@ import Table, { LinkCell, SizeCell } from 'components/common/NewTable'; import { ColumnDef } from '@tanstack/react-table'; import { clusterBrokerPath } from 'lib/paths'; +const NA = 'N/A'; + const BrokersList: React.FC = () => { const navigate = useNavigate(); const { clusterName } = useAppParams<{ clusterName: ClusterName }>(); @@ -29,14 +31,26 @@ const BrokersList: React.FC = () => { } = clusterStats; const rows = React.useMemo(() => { - if (!diskUsage) return []; + let brokersResource; + if (!diskUsage || !diskUsage?.length) { + brokersResource = + brokers?.map((broker) => { + return { + brokerId: broker.id, + segmentSize: NA, + segmentCount: NA, + }; + }) || []; + } else { + brokersResource = diskUsage; + } - return diskUsage.map(({ brokerId, segmentSize, segmentCount }) => { + return brokersResource.map(({ brokerId, segmentSize, segmentCount }) => { const broker = brokers?.find(({ id }) => id === brokerId); return { brokerId, - size: segmentSize, - count: segmentCount, + size: segmentSize || NA, + count: segmentCount || NA, port: broker?.port, host: broker?.host, }; @@ -55,7 +69,24 @@ const BrokersList: React.FC = () => { /> ), }, - { header: 'Segment Size', accessorKey: 'size', cell: SizeCell }, + { + header: 'Segment Size', + accessorKey: 'size', + // eslint-disable-next-line react/no-unstable-nested-components + cell: ({ getValue, table, cell, column, renderValue, row }) => + getValue() === NA ? ( + NA + ) : ( + <SizeCell + table={table} + column={column} + row={row} + cell={cell} + getValue={getValue} + renderValue={renderValue} + /> + ), + }, { header: 'Segment Count', accessorKey: 'count' }, { header: 'Port', accessorKey: 'port' }, { header: 'Host', accessorKey: 'host' }, diff --git a/kafka-ui-react-app/src/components/Brokers/BrokersList/__test__/BrokersList.spec.tsx b/kafka-ui-react-app/src/components/Brokers/BrokersList/__test__/BrokersList.spec.tsx index b11d477b600..cf36c64059d 100644 --- a/kafka-ui-react-app/src/components/Brokers/BrokersList/__test__/BrokersList.spec.tsx +++ b/kafka-ui-react-app/src/components/Brokers/BrokersList/__test__/BrokersList.spec.tsx @@ -134,12 +134,36 @@ describe('BrokersList Component', () => { })); }); - it('renders empty table', async () => { + describe('when it has no brokers', () => { + beforeEach(() => { + (useBrokers as jest.Mock).mockImplementation(() => ({ + data: [], + })); + }); + + it('renders empty table', async () => { + renderComponent(); + expect(screen.getByRole('table')).toBeInTheDocument(); + expect( + screen.getByRole('row', { name: 'Disk usage data not available' }) + ).toBeInTheDocument(); + }); + }); + + it('renders list of all brokers', async () => { renderComponent(); expect(screen.getByRole('table')).toBeInTheDocument(); - expect( - screen.getByRole('row', { name: 'Disk usage data not available' }) - ).toBeInTheDocument(); + expect(screen.getAllByRole('row').length).toEqual(3); + }); + it('opens broker when row clicked', async () => { + renderComponent(); + await userEvent.click(screen.getByRole('cell', { name: '1' })); + + await waitFor(() => + expect(mockedUsedNavigate).toBeCalledWith( + clusterBrokerPath(clusterName, '1') + ) + ); }); }); });
null
val
val
2022-12-07T10:09:14
"2022-10-13T06:47:50Z"
iliax
train
provectus/kafka-ui/2474_2903
provectus/kafka-ui
provectus/kafka-ui/2474
provectus/kafka-ui/2903
[ "connected" ]
fdde380072f8a269607f6ac0ef683813f31a0141
e07ce28b2c021e4e572f639be7ff6b7c10bf9cb2
[ "Hello there OpenCoderX! πŸ‘‹\n\nThank you and congratulations πŸŽ‰ for opening your very first issue in this project! πŸ’–\n\nIn case you want to claim this issue, please comment down below! We will try to get back to you as soon as we can. πŸ‘€", "relates to https://github.com/provectus/kafka-ui/issues/2738, should be fixed when https://github.com/provectus/kafka-ui/pull/2903 merged" ]
[ "please check the Linter warnings", "i think there with the new version of user event , you can automatically await the userEvent ", "fixed", "fixed", "why are we deleting the empty table case ?", "fixed" ]
"2022-11-08T11:31:30Z"
[ "type/bug", "scope/frontend", "status/accepted" ]
Cluster with 4 nodes but only 3 are listed under brokers section
- Cluster with 4 nodes but only 3 are listed under brokers section. - In the call to /stats there is only disk usage returned for 3 brokers, the broker that is not returned in disk stats is the missing broker - There are 4 brokers returned in the json response from GET http://localhost:8080/api/clusters/sensor/brokers **Set up** App version: [UI for Apache Kafka](http://kube1.semanticfact.com:8080/) vv0.4.0( [521ba0c](https://github.com/provectus/kafka-ui/commit/521ba0c) ) Kafka version: 3.2.0 in kraft mode **Screenshots** https://cdn.discordapp.com/attachments/897805035122077719/1010777595882782750/unknown.png
[ "kafka-ui-react-app/src/components/Brokers/BrokersList/BrokersList.tsx", "kafka-ui-react-app/src/components/Brokers/BrokersList/__test__/BrokersList.spec.tsx" ]
[ "kafka-ui-react-app/src/components/Brokers/BrokersList/BrokersList.tsx", "kafka-ui-react-app/src/components/Brokers/BrokersList/__test__/BrokersList.spec.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Brokers/BrokersList/BrokersList.tsx b/kafka-ui-react-app/src/components/Brokers/BrokersList/BrokersList.tsx index 3fce89fbae9..d74c1bcb700 100644 --- a/kafka-ui-react-app/src/components/Brokers/BrokersList/BrokersList.tsx +++ b/kafka-ui-react-app/src/components/Brokers/BrokersList/BrokersList.tsx @@ -10,6 +10,8 @@ import Table, { LinkCell, SizeCell } from 'components/common/NewTable'; import { ColumnDef } from '@tanstack/react-table'; import { clusterBrokerPath } from 'lib/paths'; +const NA = 'N/A'; + const BrokersList: React.FC = () => { const navigate = useNavigate(); const { clusterName } = useAppParams<{ clusterName: ClusterName }>(); @@ -29,14 +31,26 @@ const BrokersList: React.FC = () => { } = clusterStats; const rows = React.useMemo(() => { - if (!diskUsage) return []; + let brokersResource; + if (!diskUsage || !diskUsage?.length) { + brokersResource = + brokers?.map((broker) => { + return { + brokerId: broker.id, + segmentSize: NA, + segmentCount: NA, + }; + }) || []; + } else { + brokersResource = diskUsage; + } - return diskUsage.map(({ brokerId, segmentSize, segmentCount }) => { + return brokersResource.map(({ brokerId, segmentSize, segmentCount }) => { const broker = brokers?.find(({ id }) => id === brokerId); return { brokerId, - size: segmentSize, - count: segmentCount, + size: segmentSize || NA, + count: segmentCount || NA, port: broker?.port, host: broker?.host, }; @@ -55,7 +69,24 @@ const BrokersList: React.FC = () => { /> ), }, - { header: 'Segment Size', accessorKey: 'size', cell: SizeCell }, + { + header: 'Segment Size', + accessorKey: 'size', + // eslint-disable-next-line react/no-unstable-nested-components + cell: ({ getValue, table, cell, column, renderValue, row }) => + getValue() === NA ? ( + NA + ) : ( + <SizeCell + table={table} + column={column} + row={row} + cell={cell} + getValue={getValue} + renderValue={renderValue} + /> + ), + }, { header: 'Segment Count', accessorKey: 'count' }, { header: 'Port', accessorKey: 'port' }, { header: 'Host', accessorKey: 'host' }, diff --git a/kafka-ui-react-app/src/components/Brokers/BrokersList/__test__/BrokersList.spec.tsx b/kafka-ui-react-app/src/components/Brokers/BrokersList/__test__/BrokersList.spec.tsx index b11d477b600..cf36c64059d 100644 --- a/kafka-ui-react-app/src/components/Brokers/BrokersList/__test__/BrokersList.spec.tsx +++ b/kafka-ui-react-app/src/components/Brokers/BrokersList/__test__/BrokersList.spec.tsx @@ -134,12 +134,36 @@ describe('BrokersList Component', () => { })); }); - it('renders empty table', async () => { + describe('when it has no brokers', () => { + beforeEach(() => { + (useBrokers as jest.Mock).mockImplementation(() => ({ + data: [], + })); + }); + + it('renders empty table', async () => { + renderComponent(); + expect(screen.getByRole('table')).toBeInTheDocument(); + expect( + screen.getByRole('row', { name: 'Disk usage data not available' }) + ).toBeInTheDocument(); + }); + }); + + it('renders list of all brokers', async () => { renderComponent(); expect(screen.getByRole('table')).toBeInTheDocument(); - expect( - screen.getByRole('row', { name: 'Disk usage data not available' }) - ).toBeInTheDocument(); + expect(screen.getAllByRole('row').length).toEqual(3); + }); + it('opens broker when row clicked', async () => { + renderComponent(); + await userEvent.click(screen.getByRole('cell', { name: '1' })); + + await waitFor(() => + expect(mockedUsedNavigate).toBeCalledWith( + clusterBrokerPath(clusterName, '1') + ) + ); }); }); });
null
val
val
2022-12-07T10:09:14
"2022-08-21T20:45:01Z"
OpenCoderX
train
provectus/kafka-ui/2889_2904
provectus/kafka-ui
provectus/kafka-ui/2889
provectus/kafka-ui/2904
[ "connected" ]
8e35fd628abf2d08c56d7a3003c6cf4c43ec34c4
fce4c23276ebd798f182a2af7150cabffb2d63db
[]
[ "remove commented line, pls", "locator is too abstract, let's try to specify it", "var name should start from low letter if it's not constant", "this class looks overloaded with a lot of enums\ncreated pls kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/enums/ENUM_CLASS_FOR_EVERY_SINGLE_ENUM_SEPARATELY.java", "let's place those methods above inner class", "isDeleteCustomParameterBtn() -> isDeleteCustomParameterButtonEnabled", "Fixed", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "isTrue() or isFalse() missed, assertion is incomplete", "really don't understand why do we need to assert this and 296 line\ni believe we should leave 291 and 297 lines assertion, but we don't need to do this softly", "Fixed.", "Fixed." ]
"2022-11-08T11:49:52Z"
[ "scope/QA", "scope/AQA" ]
[e2e] Checking requiredness of Custom parameters within 'Create new Topic'
Autotest implementation for: https://app.qase.io/case/KAFKAUI-6 Description: Checking requiredness of Custom parameters' within 'Create new Topic' Pre-conditions: Pre-conditions Login to Kafka-ui application Open the 'Local' section Select the 'Topics' Redirect to 'Create new Topic' page with click on '+ Add a Topic' button from 'All Topics' page Post-conditions: Empty value Steps: 1 Click on '+ Add Custom Parameter' button 2 Click on 'Custom Parameter *' 3 Select the Custom Parameter value 4 Remove the "Value" field's data Expected result: 1 Should appear 'Custom Parameter' dropdown field, 'Value' input field and 'x' delete button 2 Should open Custom parameters list 3 Select the Custom Parameter value 4 Should appear "Value is required" validation message for empty 'Value'
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Topic.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicCreateEditForm.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Topic.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicCreateEditForm.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/enums/CleanupPolicyValue.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/enums/CustomParameterType.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/enums/MaxSizeOnDisk.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Topic.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Topic.java index d9a306a4f57..597529a92a4 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Topic.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/models/Topic.java @@ -1,6 +1,8 @@ package com.provectus.kafka.ui.models; -import com.provectus.kafka.ui.pages.topic.TopicCreateEditForm; +import com.provectus.kafka.ui.pages.topic.enums.CleanupPolicyValue; +import com.provectus.kafka.ui.pages.topic.enums.CustomParameterType; +import com.provectus.kafka.ui.pages.topic.enums.MaxSizeOnDisk; import lombok.Data; import lombok.experimental.Accessors; @@ -8,7 +10,8 @@ @Accessors(chain = true) public class Topic { private String name, timeToRetainData, maxMessageBytes, messageKey, messageContent, - partitions, customParameter; - private TopicCreateEditForm.CleanupPolicyValue cleanupPolicyValue; - private TopicCreateEditForm.MaxSizeOnDisk maxSizeOnDisk; -} \ No newline at end of file + partitions, customParameterValue; + private CustomParameterType customParameterType; + private CleanupPolicyValue cleanupPolicyValue; + private MaxSizeOnDisk maxSizeOnDisk; +} diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicCreateEditForm.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicCreateEditForm.java index 9a97e6dba32..a3f723164cb 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicCreateEditForm.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicCreateEditForm.java @@ -1,290 +1,288 @@ package com.provectus.kafka.ui.pages.topic; +import static com.codeborne.selenide.Selenide.$; +import static com.codeborne.selenide.Selenide.$$; +import static com.codeborne.selenide.Selenide.$x; +import static com.provectus.kafka.ui.utilities.WebUtils.clearByKeyboard; +import static com.provectus.kafka.ui.utilities.WebUtils.clickByJavaScript; +import static com.provectus.kafka.ui.utilities.WebUtils.isEnabled; +import static com.provectus.kafka.ui.utilities.WebUtils.isVisible; +import static org.assertj.core.api.Assertions.assertThat; + import com.codeborne.selenide.ClickOptions; import com.codeborne.selenide.Condition; import com.codeborne.selenide.ElementsCollection; import com.codeborne.selenide.SelenideElement; +import com.provectus.kafka.ui.pages.topic.enums.CleanupPolicyValue; +import com.provectus.kafka.ui.pages.topic.enums.CustomParameterType; +import com.provectus.kafka.ui.pages.topic.enums.MaxSizeOnDisk; import io.qameta.allure.Step; -import static com.codeborne.selenide.Selenide.*; -import static com.provectus.kafka.ui.utilities.WebUtils.clickByJavaScript; -import static com.provectus.kafka.ui.utilities.WebUtils.isEnabled; -import static org.assertj.core.api.Assertions.assertThat; - public class TopicCreateEditForm { - protected SelenideElement loadingSpinner = $x("//*[contains(text(),'Loading')]"); - protected SelenideElement timeToRetainField = $x("//input[@id='timeToRetain']"); - protected SelenideElement partitionsField = $x("//input[@name='partitions']"); - protected SelenideElement nameField = $x("//input[@name='name']"); - protected SelenideElement maxMessageBytesField = $x("//input[@name='maxMessageBytes']"); - protected SelenideElement minInSyncReplicasField = $x("//input[@name='minInSyncReplicas']"); - protected SelenideElement cleanUpPolicyDdl = $x("//ul[@id='topicFormCleanupPolicy']"); - protected SelenideElement maxSizeOnDiscDdl = $x("//ul[@id='topicFormRetentionBytes']"); - protected SelenideElement createTopicBtn = $x("//button[@type='submit']"); - protected String ddlElementLocator = "//li[@value='%s']"; - - @Step - public TopicCreateEditForm waitUntilScreenReady(){ - loadingSpinner.shouldBe(Condition.disappear); - nameField.shouldBe(Condition.visible); - return this; - } - - @Step - public TopicCreateEditForm setTopicName(String topicName) { - nameField.setValue(topicName); - return this; - } - - @Step - public TopicCreateEditForm setMinInsyncReplicas(Integer minInsyncReplicas) { - minInSyncReplicasField.setValue(minInsyncReplicas.toString()); - return this; - } - - @Step - public TopicCreateEditForm setTimeToRetainDataInMs(Long ms) { - timeToRetainField.setValue(ms.toString()); - return this; - } - - @Step - public TopicCreateEditForm setTimeToRetainDataInMs(String ms) { - timeToRetainField.setValue(ms); - return this; - } - - @Step - public TopicCreateEditForm setMaxSizeOnDiskInGB(MaxSizeOnDisk MaxSizeOnDisk) { - maxSizeOnDiscDdl.shouldBe(Condition.visible).click(); - $x(String.format(ddlElementLocator, MaxSizeOnDisk.getOptionValue())).shouldBe(Condition.visible).click(); - return this; - } - - @Step - public TopicCreateEditForm setMaxMessageBytes(Long bytes) { - maxMessageBytesField.setValue(bytes.toString()); - return this; - } - - @Step - public TopicCreateEditForm setMaxMessageBytes(String bytes) { - return setMaxMessageBytes(Long.parseLong(bytes)); - } - - @Step - public TopicCreateEditForm setPartitions(String partitions){ - partitionsField.setValue(partitions); - return this; - } - - @Step - public TopicCreateEditForm setTimeToRetainDataInMsUsingButtons(String value) { - timeToRetainField - .parent() - .parent() - .$$("button") - .find(Condition.exactText(value)) - .click(); - return this; - } - - @Step - public TopicCreateEditForm selectCleanupPolicy(CleanupPolicyValue cleanupPolicyOptionValue) { - cleanUpPolicyDdl.shouldBe(Condition.visible).click(); - $x(String.format(ddlElementLocator,cleanupPolicyOptionValue.getOptionValue())).shouldBe(Condition.visible).click(); - return this; - } - - @Step - public TopicCreateEditForm selectRetentionBytes(String visibleValue) { - return selectFromDropDownByVisibleText("retentionBytes", visibleValue); - } - - @Step - public TopicCreateEditForm selectRetentionBytes(Long optionValue) { - return selectFromDropDownByOptionValue("retentionBytes", optionValue.toString()); - } - - @Step - public TopicCreateEditForm clickCreateTopicBtn() { - clickByJavaScript(createTopicBtn); - return this; + protected SelenideElement loadingSpinner = $x("//*[contains(text(),'Loading')]"); + protected SelenideElement timeToRetainField = $x("//input[@id='timeToRetain']"); + protected SelenideElement partitionsField = $x("//input[@name='partitions']"); + protected SelenideElement nameField = $x("//input[@name='name']"); + protected SelenideElement maxMessageBytesField = $x("//input[@name='maxMessageBytes']"); + protected SelenideElement minInSyncReplicasField = $x("//input[@name='minInSyncReplicas']"); + protected SelenideElement cleanUpPolicyDdl = $x("//ul[@id='topicFormCleanupPolicy']"); + protected SelenideElement maxSizeOnDiscDdl = $x("//ul[@id='topicFormRetentionBytes']"); + protected SelenideElement customParameterDdl = $x("//ul[contains(@name,'customParams')]"); + protected SelenideElement createTopicBtn = $x("//button[@type='submit']"); + protected SelenideElement deleteCustomParameterBtn = $x("//span[contains(@title,'Delete customParam')]"); + protected SelenideElement addCustomParameterTypeBtn = $x("//button[contains(text(),'Add Custom Parameter')]"); + protected SelenideElement customParameterValueField = $x("//input[@placeholder='Value']"); + protected SelenideElement validationCustomParameterValueMsg = $x("//p[contains(text(),'Value is required')]"); + protected String ddlElementLocator = "//li[@value='%s']"; + + @Step + public TopicCreateEditForm waitUntilScreenReady() { + loadingSpinner.shouldBe(Condition.disappear); + nameField.shouldBe(Condition.visible); + return this; + } + + public boolean isCreateTopicButtonEnabled() { + return isEnabled(createTopicBtn); + } + + public boolean isDeleteCustomParameterButtonEnabled() { + return isEnabled(deleteCustomParameterBtn); + } + + @Step + public TopicCreateEditForm setTopicName(String topicName) { + nameField.setValue(topicName); + return this; + } + + @Step + public TopicCreateEditForm setMinInsyncReplicas(Integer minInsyncReplicas) { + minInSyncReplicasField.setValue(minInsyncReplicas.toString()); + return this; + } + + @Step + public TopicCreateEditForm setTimeToRetainDataInMs(Long ms) { + timeToRetainField.setValue(ms.toString()); + return this; + } + + @Step + public TopicCreateEditForm setTimeToRetainDataInMs(String ms) { + timeToRetainField.setValue(ms); + return this; + } + + @Step + public TopicCreateEditForm setMaxSizeOnDiskInGB(MaxSizeOnDisk MaxSizeOnDisk) { + maxSizeOnDiscDdl.shouldBe(Condition.visible).click(); + $x(String.format(ddlElementLocator, MaxSizeOnDisk.getOptionValue())).shouldBe(Condition.visible).click(); + return this; + } + + @Step + public TopicCreateEditForm clickAddCustomParameterTypeButton() { + addCustomParameterTypeBtn.click(); + return this; + } + + @Step + public TopicCreateEditForm setCustomParameterType(CustomParameterType customParameterType) { + customParameterDdl.shouldBe(Condition.visible).click(); + $x(String.format(ddlElementLocator, customParameterType.getOptionValue())).shouldBe(Condition.visible).click(); + return this; + } + + @Step + public TopicCreateEditForm clearCustomParameterValue() { + clearByKeyboard(customParameterValueField); + return this; + } + + @Step + public TopicCreateEditForm setMaxMessageBytes(Long bytes) { + maxMessageBytesField.setValue(bytes.toString()); + return this; + } + + @Step + public TopicCreateEditForm setMaxMessageBytes(String bytes) { + return setMaxMessageBytes(Long.parseLong(bytes)); + } + + @Step + public TopicCreateEditForm setPartitions(String partitions) { + partitionsField.setValue(partitions); + return this; + } + + @Step + public TopicCreateEditForm setTimeToRetainDataInMsUsingButtons(String value) { + timeToRetainField + .parent() + .parent() + .$$("button") + .find(Condition.exactText(value)) + .click(); + return this; + } + + @Step + public TopicCreateEditForm selectCleanupPolicy(CleanupPolicyValue cleanupPolicyOptionValue) { + cleanUpPolicyDdl.shouldBe(Condition.visible).click(); + $x(String.format(ddlElementLocator, cleanupPolicyOptionValue.getOptionValue())).shouldBe(Condition.visible).click(); + return this; + } + + @Step + public TopicCreateEditForm selectRetentionBytes(String visibleValue) { + return selectFromDropDownByVisibleText("retentionBytes", visibleValue); + } + + @Step + public TopicCreateEditForm selectRetentionBytes(Long optionValue) { + return selectFromDropDownByOptionValue("retentionBytes", optionValue.toString()); + } + + @Step + public TopicCreateEditForm clickCreateTopicBtn() { + clickByJavaScript(createTopicBtn); + return this; + } + + @Step + public TopicCreateEditForm addCustomParameter(String customParameterName, + String customParameterValue) { + ElementsCollection customParametersElements = + $$("ul[role=listbox][name^=customParams][name$=name]"); + KafkaUISelectElement kafkaUISelectElement = null; + if (customParametersElements.size() == 1) { + if ("Select".equals(customParametersElements.first().getText())) { + kafkaUISelectElement = new KafkaUISelectElement(customParametersElements.first()); + } + } else { + $$("button") + .find(Condition.exactText("Add Custom Parameter")) + .click(); + customParametersElements = $$("ul[role=listbox][name^=customParams][name$=name]"); + kafkaUISelectElement = new KafkaUISelectElement(customParametersElements.last()); } - - @Step - public TopicCreateEditForm addCustomParameter(String customParameterName, - String customParameterValue) { - ElementsCollection customParametersElements = - $$("ul[role=listbox][name^=customParams][name$=name]"); - KafkaUISelectElement kafkaUISelectElement = null; - if (customParametersElements.size() == 1) { - if ("Select".equals(customParametersElements.first().getText())) { - kafkaUISelectElement = new KafkaUISelectElement(customParametersElements.first()); - } - } else { - $$("button") - .find(Condition.exactText("Add Custom Parameter")) - .click(); - customParametersElements = $$("ul[role=listbox][name^=customParams][name$=name]"); - kafkaUISelectElement = new KafkaUISelectElement(customParametersElements.last()); - } - if (kafkaUISelectElement != null) { - kafkaUISelectElement.selectByVisibleText(customParameterName); - } - $(String.format("input[name=\"customParams.%d.value\"]", customParametersElements.size() - 1)) - .setValue(customParameterValue); - return this; + if (kafkaUISelectElement != null) { + kafkaUISelectElement.selectByVisibleText(customParameterName); } - - @Step - public TopicCreateEditForm updateCustomParameter(String customParameterName, - String customParameterValue) { - SelenideElement selenideElement = $$("ul[role=listbox][name^=customParams][name$=name]") - .find(Condition.exactText(customParameterName)); - String name = selenideElement.getAttribute("name"); - if (name != null) { - name = name.substring(0, name.lastIndexOf(".")); - } - $(String.format("input[name^=%s]", name)).setValue(customParameterValue); - return this; + $(String.format("input[name=\"customParams.%d.value\"]", customParametersElements.size() - 1)) + .setValue(customParameterValue); + return this; + } + + @Step + public TopicCreateEditForm updateCustomParameter(String customParameterName, + String customParameterValue) { + SelenideElement selenideElement = $$("ul[role=listbox][name^=customParams][name$=name]") + .find(Condition.exactText(customParameterName)); + String name = selenideElement.getAttribute("name"); + if (name != null) { + name = name.substring(0, name.lastIndexOf(".")); } - - @Step - public TopicCreateEditForm cleanupPolicyIs(String value) { - String cleanupPolicy = new KafkaUISelectElement("cleanupPolicy") - .getCurrentValue(); - assertThat(cleanupPolicy) - .as("Clear policy value should be " + value) - .isEqualToIgnoringCase(value); - return this; + $(String.format("input[name^=%s]", name)).setValue(customParameterValue); + return this; + } + + @Step + public TopicCreateEditForm cleanupPolicyIs(String value) { + String cleanupPolicy = new KafkaUISelectElement("cleanupPolicy") + .getCurrentValue(); + assertThat(cleanupPolicy) + .as("Clear policy value should be " + value) + .isEqualToIgnoringCase(value); + return this; + } + + @Step + public TopicCreateEditForm timeToRetainIs(String time) { + String value = timeToRetainField.getValue(); + assertThat(value) + .as("Time to retain data (in ms) should be " + time) + .isEqualTo(time); + return this; + } + + @Step + public String getCleanupPolicy() { + return new KafkaUISelectElement("cleanupPolicy").getCurrentValue(); + } + + @Step + public String getTimeToRetain() { + return timeToRetainField.getValue(); + } + + @Step + public String getMaxSizeOnDisk() { + return new KafkaUISelectElement("retentionBytes").getCurrentValue(); + } + + @Step + public String getMaxMessageBytes() { + return maxMessageBytesField.getValue(); + } + + @Step + public boolean isValidationMessageCustomParameterValueVisible() { + return isVisible(validationCustomParameterValueMsg); + } + + @Step + public String getCustomParameterValue() { + return customParameterValueField.getValue(); + } + + private static class KafkaUISelectElement { + + private final SelenideElement selectElement; + + public KafkaUISelectElement(String selectElementName) { + this.selectElement = $("ul[role=listbox][name=" + selectElementName + "]"); } - @Step - public TopicCreateEditForm timeToRetainIs(String time) { - String value = timeToRetainField.getValue(); - assertThat(value) - .as("Time to retain data (in ms) should be " + time) - .isEqualTo(time); - return this; - } - - @Step - public String getCleanupPolicy() { - return new KafkaUISelectElement("cleanupPolicy").getCurrentValue(); - } - - @Step - public String getTimeToRetain() { - return timeToRetainField.getValue(); - } - - @Step - public String getMaxSizeOnDisk() { - return new KafkaUISelectElement("retentionBytes").getCurrentValue(); - } - - @Step - public String getMaxMessageBytes() { - return maxMessageBytesField.getValue(); - } - - - private static class KafkaUISelectElement { - - private final SelenideElement selectElement; - - public KafkaUISelectElement(String selectElementName) { - this.selectElement = $("ul[role=listbox][name=" + selectElementName + "]"); - } - - public KafkaUISelectElement(SelenideElement selectElement) { - this.selectElement = selectElement; - } - - public void selectByOptionValue(String optionValue) { - selectElement.click(); - selectElement - .$$x(".//ul/li[@role='option']") - .find(Condition.attribute("value", optionValue)) - .click(ClickOptions.usingJavaScript()); - } - - public void selectByVisibleText(String visibleText) { - selectElement.click(); - selectElement - .$$("ul>li[role=option]") - .find(Condition.exactText(visibleText)) - .click(); - } - - public String getCurrentValue() { - return selectElement.$("li").getText(); - } - } - - public enum CleanupPolicyValue { - DELETE("delete", "Delete"), - COMPACT("compact", "Compact"), - COMPACT_DELETE("compact,delete", "Compact,Delete"); - - private final String optionValue; - private final String visibleText; - - CleanupPolicyValue(String optionValue, String visibleText) { - this.optionValue = optionValue; - this.visibleText = visibleText; - } - - public String getOptionValue() { - return optionValue; - } - - public String getVisibleText() { - return visibleText; - } - } - - public enum MaxSizeOnDisk { - NOT_SET("-1", "Not Set"), - SIZE_1_GB("1073741824", "1 GB"), - SIZE_10_GB("10737418240", "10 GB"), - SIZE_20_GB("21474836480", "20 GB"), - SIZE_50_GB("53687091200", "50 GB"); - - private final String optionValue; - private final String visibleText; - - MaxSizeOnDisk(String optionValue, String visibleText) { - this.optionValue = optionValue; - this.visibleText = visibleText; - } - - public String getOptionValue() { - return optionValue; - } - - public String getVisibleText() { - return visibleText; - } + public KafkaUISelectElement(SelenideElement selectElement) { + this.selectElement = selectElement; } - public boolean isCreateTopicButtonEnabled(){ - return isEnabled(createTopicBtn); + public void selectByOptionValue(String optionValue) { + selectElement.click(); + selectElement + .$$x(".//ul/li[@role='option']") + .find(Condition.attribute("value", optionValue)) + .click(ClickOptions.usingJavaScript()); } - private TopicCreateEditForm selectFromDropDownByOptionValue(String dropDownElementName, - String optionValue) { - KafkaUISelectElement select = new KafkaUISelectElement(dropDownElementName); - select.selectByOptionValue(optionValue); - return this; + public void selectByVisibleText(String visibleText) { + selectElement.click(); + selectElement + .$$("ul>li[role=option]") + .find(Condition.exactText(visibleText)) + .click(); } - private TopicCreateEditForm selectFromDropDownByVisibleText(String dropDownElementName, - String visibleText) { - KafkaUISelectElement select = new KafkaUISelectElement(dropDownElementName); - select.selectByVisibleText(visibleText); - return this; + public String getCurrentValue() { + return selectElement.$("li").getText(); } + } + + private TopicCreateEditForm selectFromDropDownByOptionValue(String dropDownElementName, + String optionValue) { + KafkaUISelectElement select = new KafkaUISelectElement(dropDownElementName); + select.selectByOptionValue(optionValue); + return this; + } + + private TopicCreateEditForm selectFromDropDownByVisibleText(String dropDownElementName, + String visibleText) { + KafkaUISelectElement select = new KafkaUISelectElement(dropDownElementName); + select.selectByVisibleText(visibleText); + return this; + } } diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/enums/CleanupPolicyValue.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/enums/CleanupPolicyValue.java new file mode 100644 index 00000000000..86b87325a0d --- /dev/null +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/enums/CleanupPolicyValue.java @@ -0,0 +1,24 @@ +package com.provectus.kafka.ui.pages.topic.enums; + +public enum CleanupPolicyValue { + DELETE("delete", "Delete"), + COMPACT("compact", "Compact"), + COMPACT_DELETE("compact,delete", "Compact,Delete"); + + private final String optionValue; + private final String visibleText; + + CleanupPolicyValue(String optionValue, String visibleText) { + this.optionValue = optionValue; + this.visibleText = visibleText; + } + + public String getOptionValue() { + return optionValue; + } + + public String getVisibleText() { + return visibleText; + } +} + diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/enums/CustomParameterType.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/enums/CustomParameterType.java new file mode 100644 index 00000000000..bc2f3befaee --- /dev/null +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/enums/CustomParameterType.java @@ -0,0 +1,36 @@ +package com.provectus.kafka.ui.pages.topic.enums; + +public enum CustomParameterType { + COMPRESSION_TYPE("compression.type"), + DELETE_RETENTION_MS("delete.retention.ms"), + FILE_DELETE_DELAY_MS("file.delete.delay.ms"), + FLUSH_MESSAGES("flush.messages"), + FLUSH_MS("flush.ms"), + FOLLOWER_REPLICATION_THROTTLED_REPLICAS("follower.replication.throttled.replicas"), + INDEX_INTERVAL_BYTES("index.interval.bytes"), + LEADER_REPLICATION_THROTTLED_REPLICAS("leader.replication.throttled.replicas"), + MAX_COMPACTION_LAG_MS("max.compaction.lag.ms"), + MESSAGE_DOWNCONVERSION_ENABLE("message.downconversion.enable"), + MESSAGE_FORMAT_VERSION("message.format.version"), + MESSAGE_TIMESTAMP_DIFFERENCE_MAX_MS("message.timestamp.difference.max.ms"), + MESSAGE_TIMESTAMP_TYPE("message.timestamp.type"), + MIN_CLEANABLE_DIRTY_RATIO("min.cleanable.dirty.ratio"), + MIN_COMPACTION_LAG_MS("min.compaction.lag.ms"), + PREALLOCATE("preallocate"), + RETENTION_BYTES("retention.bytes"), + SEGMENT_BYTES("segment.bytes"), + SEGMENT_INDEX_BYTES("segment.index.bytes"), + SEGMENT_JITTER_MS("segment.jitter.ms"), + SEGMENT_MS("segment.ms"), + UNCLEAN_LEADER_ELECTION_ENABLE("unclean.leader.election.enable"); + + private final String optionValue; + + CustomParameterType(String optionValue) { + this.optionValue = optionValue; + } + + public String getOptionValue() { + return optionValue; + } +} diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/enums/MaxSizeOnDisk.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/enums/MaxSizeOnDisk.java new file mode 100644 index 00000000000..bdb476d1e6d --- /dev/null +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/enums/MaxSizeOnDisk.java @@ -0,0 +1,26 @@ +package com.provectus.kafka.ui.pages.topic.enums; + +public enum MaxSizeOnDisk { + NOT_SET("-1", "Not Set"), + SIZE_1_GB("1073741824", "1 GB"), + SIZE_10_GB("10737418240", "10 GB"), + SIZE_20_GB("21474836480", "20 GB"), + SIZE_50_GB("53687091200", "50 GB"); + + private final String optionValue; + private final String visibleText; + + MaxSizeOnDisk(String optionValue, String visibleText) { + this.optionValue = optionValue; + this.visibleText = visibleText; + } + + public String getOptionValue() { + return optionValue; + } + + public String getVisibleText() { + return visibleText; + } +} +
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java index d707b5acafd..aaa7a6fe8b6 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java @@ -1,9 +1,10 @@ package com.provectus.kafka.ui.tests; import static com.provectus.kafka.ui.pages.NaviSideBar.SideMenuOption.TOPICS; -import static com.provectus.kafka.ui.pages.topic.TopicCreateEditForm.CleanupPolicyValue.COMPACT; -import static com.provectus.kafka.ui.pages.topic.TopicCreateEditForm.CleanupPolicyValue.DELETE; -import static com.provectus.kafka.ui.pages.topic.TopicCreateEditForm.MaxSizeOnDisk.SIZE_20_GB; +import static com.provectus.kafka.ui.pages.topic.enums.CleanupPolicyValue.COMPACT; +import static com.provectus.kafka.ui.pages.topic.enums.CleanupPolicyValue.DELETE; +import static com.provectus.kafka.ui.pages.topic.enums.CustomParameterType.COMPRESSION_TYPE; +import static com.provectus.kafka.ui.pages.topic.enums.MaxSizeOnDisk.SIZE_20_GB; import static com.provectus.kafka.ui.settings.Source.CLUSTER_NAME; import static com.provectus.kafka.ui.utilities.FileUtils.fileToString; import static org.apache.commons.lang.RandomStringUtils.randomAlphabetic; @@ -35,6 +36,8 @@ public class TopicTests extends BaseTest { private static final Topic TOPIC_TO_CREATE = new Topic() .setName("new-topic-"+ randomAlphabetic(5)) .setPartitions("1") + .setCustomParameterType(COMPRESSION_TYPE) + .setCustomParameterValue("producer") .setCleanupPolicyValue(DELETE); private static final Topic TOPIC_FOR_UPDATE = new Topic() .setName("topic-to-update-" + randomAlphabetic(5)) @@ -268,6 +271,30 @@ void checkTopicCreatePossibility() { assertThat(topicCreateEditForm.isCreateTopicButtonEnabled()).as("isCreateTopicButtonEnabled()").isTrue(); } + @DisplayName("Checking requiredness of Custom parameters within 'Create new Topic'") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(6) + @Test + void checkCustomParametersWithinCreateNewTopic() { + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .clickAddTopicBtn(); + topicCreateEditForm + .waitUntilScreenReady() + .setTopicName(TOPIC_TO_CREATE.getName()) + .clickAddCustomParameterTypeButton() + .setCustomParameterType(TOPIC_TO_CREATE.getCustomParameterType()); + assertThat(topicCreateEditForm.isDeleteCustomParameterButtonEnabled()).as("isDeleteCustomParameterButtonEnabled()") + .isTrue(); + topicCreateEditForm + .clearCustomParameterValue(); + assertThat(topicCreateEditForm.isValidationMessageCustomParameterValueVisible()) + .as("isValidationMessageCustomParameterValueVisible()").isTrue(); + } + @AfterAll public void afterAll() { TOPIC_LIST.forEach(topic -> apiHelper.deleteTopic(CLUSTER_NAME, topic.getName()));
train
val
2022-11-08T12:28:20
"2022-11-07T09:35:41Z"
ArthurNiedial
train
provectus/kafka-ui/2923_2924
provectus/kafka-ui
provectus/kafka-ui/2923
provectus/kafka-ui/2924
[ "connected" ]
8bb9f6365823ed8ad181f68ec3a37e6652c4e39e
93af20c46cacef03fbd06eceebe3a6caffda7b0d
[ "PR created https://github.com/provectus/kafka-ui/pull/2924 for that issue." ]
[]
"2022-11-11T11:31:25Z"
[ "type/enhancement", "status/accepted", "scope/k8s" ]
Add posibility to define scheme as HTTPS in readiness and liveness probes
**Describe the bug** <!--(A clear and concise description of what the bug is.)--> If we enable SSL in kafka-ui backend, and deploy into k8s using the helm chart, the deployment creates the livenessProbe and readinessProbe configuration with the default value of HTTP. https://kubernetes.io/docs/tasks/configure-pod-container/configure-liveness-readiness-startup-probes/#http-probes In the current Deployment.yaml there is not way to specify the scheme field for probes **Set up** <!-- How do you run the app? Please provide as much info as possible: 1. App version (docker image version or check commit hash in the top left corner in UI) 2. Helm chart version, if you use one 3. Any IAAC configs We might close the issue without further explanation if you don't provide such information. --> **Steps to Reproduce** Steps to reproduce the behavior: 1. Enable SSL in Kafka-ui. (SERVER_SSL_ENABLED=true, SERVER_SSL_KEY_STORE='myserver.keystore', SERVER_SSL_KEY_STORE_PASSWORD='mypassword' 2. Deploy the application in kubernetes using the helm chart. 3. Kubernetes restart the pods, since there is not valid certificate for that operation. **Expected behavior** Make able to configure the deployment telling to kubernetes to skipping the certificate verification.
[ "charts/kafka-ui/Chart.yaml", "charts/kafka-ui/templates/deployment.yaml", "charts/kafka-ui/values.yaml" ]
[ "charts/kafka-ui/Chart.yaml", "charts/kafka-ui/templates/deployment.yaml", "charts/kafka-ui/values.yaml" ]
[]
diff --git a/charts/kafka-ui/Chart.yaml b/charts/kafka-ui/Chart.yaml index 1ecf0eeecea..2ae296b4ee6 100644 --- a/charts/kafka-ui/Chart.yaml +++ b/charts/kafka-ui/Chart.yaml @@ -2,6 +2,6 @@ apiVersion: v2 name: kafka-ui description: A Helm chart for kafka-UI type: application -version: 0.4.5 +version: 0.4.6 appVersion: v0.4.0 icon: https://github.com/provectus/kafka-ui/raw/master/documentation/images/kafka-ui-logo.png diff --git a/charts/kafka-ui/templates/deployment.yaml b/charts/kafka-ui/templates/deployment.yaml index a253fc6df51..219e9b44061 100644 --- a/charts/kafka-ui/templates/deployment.yaml +++ b/charts/kafka-ui/templates/deployment.yaml @@ -87,6 +87,9 @@ spec: {{- $contextPath := .Values.envs.config.SERVER_SERVLET_CONTEXT_PATH | default "" | printf "%s/actuator/health" | urlParse }} path: {{ get $contextPath "path" }} port: http + {{- if .Values.probes.useHttpsScheme }} + scheme: HTTPS + {{- end }} initialDelaySeconds: 60 periodSeconds: 30 timeoutSeconds: 10 @@ -95,6 +98,9 @@ spec: {{- $contextPath := .Values.envs.config.SERVER_SERVLET_CONTEXT_PATH | default "" | printf "%s/actuator/health" | urlParse }} path: {{ get $contextPath "path" }} port: http + {{- if .Values.probes.useHttpsScheme }} + scheme: HTTPS + {{- end }} initialDelaySeconds: 60 periodSeconds: 30 timeoutSeconds: 10 diff --git a/charts/kafka-ui/values.yaml b/charts/kafka-ui/values.yaml index 35e3ad3dd59..dce32059e14 100644 --- a/charts/kafka-ui/values.yaml +++ b/charts/kafka-ui/values.yaml @@ -73,6 +73,11 @@ podLabels: {} ## annotations: {} +## Set field schema as HTTPS for readines and liveness probe +## +probes: + useHttpsScheme: false + podSecurityContext: {} # fsGroup: 2000
null
train
val
2022-11-17T12:17:28
"2022-11-11T11:00:28Z"
joseacl
train
provectus/kafka-ui/2914_2925
provectus/kafka-ui
provectus/kafka-ui/2914
provectus/kafka-ui/2925
[ "connected" ]
97aa114696b59167837886f554f6062094619fb5
6de731778bf9e2884f5f50a464d4f399055de81e
[ "Hello there anezboretskiy! πŸ‘‹\n\nThank you and congratulations πŸŽ‰ for opening your very first issue in this project! πŸ’–\n\nIn case you want to claim this issue, please comment down below! We will try to get back to you as soon as we can. πŸ‘€" ]
[ " public List<SelenideElement> getAllVisibleElements() {\n List<SelenideElement> visibleElements = new ArrayList<>(getVisibleColumnHeaders());\n visibleElements.addAll(Arrays.asList(searchField, addTopicBtn, tableGrid));\n visibleElements.addAll(getActionButtons());\n return visibleElements;", "Refactored " ]
"2022-11-11T12:28:43Z"
[ "scope/QA", "scope/AQA" ]
[e2e]Checking Topics section within Kafka-ui Application
**Autotest implementation for:** https://app.qase.io/case/KAFKAUI-2 **Description:** Checking 'All Topics' page content **Pre-conditions:** - Login to Kafka-ui application - Open the Local cluster in a left **Post-conditions:** Empty value **Steps:** 1. Select Topics within Local section **Expected result:** Should open 'All Topics' page with following information: - Search input field; - 'Show Internal Topics' radio-button; - '+ Add a Topic' button; - Existing Topics' list; - Filters for 'Topic Name', 'Total Partitions', 'Out of sync replicas', 'Replication Factor', 'Number of messages', 'Size'; - Pagination;
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/TopPanel.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/TopPanel.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/SmokeTests.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/connectors/ConnectorsTests.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/schemas/SchemasTests.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicActionsTests.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicMessagesTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/TopPanel.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/TopPanel.java index 8101543abfc..267bb4a9b21 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/TopPanel.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/TopPanel.java @@ -1,13 +1,12 @@ package com.provectus.kafka.ui.pages; -import com.codeborne.selenide.SelenideElement; +import static com.codeborne.selenide.Selenide.$x; +import com.codeborne.selenide.SelenideElement; import java.util.Arrays; import java.util.List; -import static com.codeborne.selenide.Selenide.$x; - -public class TopPanel { +public class TopPanel extends BasePage{ protected SelenideElement kafkaLogo = $x("//a[contains(text(),'UI for Apache Kafka')]"); protected SelenideElement kafkaVersion = $x("//a[@title='Current commit']"); protected SelenideElement logOutBtn = $x("//button[contains(text(),'Log out')]"); diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java index 61a04115fb4..9ac485c26b7 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java @@ -3,6 +3,7 @@ import static com.codeborne.selenide.Selenide.$; import static com.codeborne.selenide.Selenide.$$x; import static com.codeborne.selenide.Selenide.$x; +import static org.apache.commons.lang.math.RandomUtils.nextInt; import com.codeborne.selenide.CollectionCondition; import com.codeborne.selenide.Condition; @@ -142,7 +143,7 @@ public TopicDetails.MessageGridItem getMessage(int offset) { @Step public TopicDetails.MessageGridItem getRandomMessage() { - return getMessage(initItems().size() - 1); + return getMessage(nextInt(initItems().size() - 1)); } public static class MessageGridItem extends BasePage { diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java index 34e997152ff..cd61bd72721 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java @@ -7,6 +7,11 @@ import com.provectus.kafka.ui.pages.BasePage; import com.provectus.kafka.ui.utilities.WaitUtils; import io.qameta.allure.Step; +import java.util.ArrayList; +import java.util.Arrays; +import java.util.List; +import java.util.stream.Collectors; +import java.util.stream.Stream; import lombok.experimental.ExtensionMethod; @ExtensionMethod(WaitUtils.class) @@ -14,6 +19,10 @@ public class TopicsList extends BasePage { protected SelenideElement topicListHeader = $x("//h1[text()='Topics']"); protected SelenideElement addTopicBtn = $x("//button[normalize-space(text()) ='Add a Topic']"); + protected SelenideElement searchField = $x("//input[@placeholder='Search by Topic Name']"); + protected SelenideElement showInternalRadioBtn = $x("//input[@name='ShowInternalTopics']"); + protected String сolumnHeaderLocator = "//table//tr/th/div[text()='%s']"; + protected String actionButtonLocator = "//button[text()='%s']"; @Step public TopicsList waitUntilScreenReady() { @@ -39,4 +48,37 @@ public TopicsList openTopic(String topicName) { getTableElement(topicName).shouldBe(Condition.enabled).click(); return this; } + + private List<SelenideElement> getActionButtons() { + return Stream.of("Delete selected topics", "Copy selected topic", "Purge messages of selected topics") + .map(name -> $x(String.format(actionButtonLocator, name))) + .collect(Collectors.toList()); + } + + private List<SelenideElement> getVisibleColumnHeaders() { + return Stream.of("Replication Factor","Number of messages","Topic Name", "Partitions", "Out of sync replicas", "Size") + .map(name -> $x(String.format(сolumnHeaderLocator, name))) + .collect(Collectors.toList()); + } + + private List<SelenideElement> getEnabledColumnHeaders(){ + return Stream.of("Topic Name", "Partitions", "Out of sync replicas", "Size") + .map(name -> $x(String.format(сolumnHeaderLocator, name))) + .collect(Collectors.toList()); + } + + @Step + public List<SelenideElement> getAllVisibleElements() { + List<SelenideElement> visibleElements = new ArrayList<>(getVisibleColumnHeaders()); + visibleElements.addAll(Arrays.asList(searchField, addTopicBtn, tableGrid)); + visibleElements.addAll(getActionButtons()); + return visibleElements; + } + + @Step + public List<SelenideElement> getAllEnabledElements() { + List<SelenideElement> enabledElements = new ArrayList<>(getEnabledColumnHeaders()); + enabledElements.addAll(Arrays.asList(searchField, showInternalRadioBtn,addTopicBtn)); + return enabledElements; + } }
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/SmokeTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/SmokeTests.java similarity index 97% rename from kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/SmokeTests.java rename to kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/SmokeTests.java index e1e6927c432..ed7d901fe4b 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/SmokeTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/SmokeTests.java @@ -1,4 +1,4 @@ -package com.provectus.kafka.ui; +package com.provectus.kafka.ui.suite; import com.codeborne.selenide.Condition; import com.provectus.kafka.ui.base.BaseTest; @@ -29,4 +29,4 @@ public void checkBasePageElements(){ .as(element.getSearchCriteria() + " isEnabled()").isTrue()); softly.assertAll(); } -} \ No newline at end of file +} diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/ConnectorsTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/connectors/ConnectorsTests.java similarity index 99% rename from kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/ConnectorsTests.java rename to kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/connectors/ConnectorsTests.java index 505cb6bcf00..5e8ed9157ab 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/ConnectorsTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/connectors/ConnectorsTests.java @@ -1,4 +1,4 @@ -package com.provectus.kafka.ui.tests; +package com.provectus.kafka.ui.suite.connectors; import static com.provectus.kafka.ui.pages.BasePage.AlertHeader.SUCCESS; import static com.provectus.kafka.ui.pages.NaviSideBar.SideMenuOption.KAFKA_CONNECT; diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/SchemasTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/schemas/SchemasTests.java similarity index 99% rename from kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/SchemasTests.java rename to kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/schemas/SchemasTests.java index 681b95bbb56..553568795dd 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/SchemasTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/schemas/SchemasTests.java @@ -1,4 +1,4 @@ -package com.provectus.kafka.ui.tests; +package com.provectus.kafka.ui.suite.schemas; import static com.provectus.kafka.ui.pages.NaviSideBar.SideMenuOption.SCHEMA_REGISTRY; import static com.provectus.kafka.ui.settings.Source.CLUSTER_NAME; diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicActionsTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicActionsTests.java new file mode 100644 index 00000000000..3bc6dc07b4a --- /dev/null +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicActionsTests.java @@ -0,0 +1,259 @@ +package com.provectus.kafka.ui.suite.topics; + +import static com.provectus.kafka.ui.pages.NaviSideBar.SideMenuOption.TOPICS; +import static com.provectus.kafka.ui.pages.topic.enums.CleanupPolicyValue.COMPACT; +import static com.provectus.kafka.ui.pages.topic.enums.CleanupPolicyValue.DELETE; +import static com.provectus.kafka.ui.pages.topic.enums.CustomParameterType.COMPRESSION_TYPE; +import static com.provectus.kafka.ui.pages.topic.enums.MaxSizeOnDisk.SIZE_20_GB; +import static com.provectus.kafka.ui.settings.Source.CLUSTER_NAME; +import static com.provectus.kafka.ui.utilities.FileUtils.fileToString; +import static org.apache.commons.lang.RandomStringUtils.randomAlphabetic; +import static org.assertj.core.api.Assertions.assertThat; + +import com.codeborne.selenide.Condition; +import com.provectus.kafka.ui.base.BaseTest; +import com.provectus.kafka.ui.models.Topic; +import com.provectus.kafka.ui.pages.topic.TopicDetails; +import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.AutomationStatus; +import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.Suite; +import com.provectus.kafka.ui.utilities.qaseIoUtils.enums.Status; +import io.qameta.allure.Issue; +import io.qase.api.annotation.CaseId; +import java.util.ArrayList; +import java.util.List; +import org.assertj.core.api.SoftAssertions; +import org.junit.jupiter.api.AfterAll; +import org.junit.jupiter.api.Assertions; +import org.junit.jupiter.api.BeforeAll; +import org.junit.jupiter.api.Disabled; +import org.junit.jupiter.api.DisplayName; +import org.junit.jupiter.api.Test; +import org.junit.jupiter.api.TestInstance; + +@TestInstance(TestInstance.Lifecycle.PER_CLASS) +public class TopicActionsTests extends BaseTest { + private static final long SUITE_ID = 2; + private static final String SUITE_TITLE = "Topics"; + private static final Topic TOPIC_TO_CREATE = new Topic() + .setName("new-topic-"+ randomAlphabetic(5)) + .setPartitions("1") + .setCustomParameterType(COMPRESSION_TYPE) + .setCustomParameterValue("producer") + .setCleanupPolicyValue(DELETE); + private static final Topic TOPIC_FOR_UPDATE = new Topic() + .setName("topic-to-update-" + randomAlphabetic(5)) + .setCleanupPolicyValue(COMPACT) + .setTimeToRetainData("604800001") + .setMaxSizeOnDisk(SIZE_20_GB) + .setMaxMessageBytes("1000020") + .setMessageKey(fileToString(System.getProperty("user.dir") + "/src/test/resources/producedkey.txt")) + .setMessageContent(fileToString(System.getProperty("user.dir") + "/src/test/resources/testData.txt")); + + private static final Topic TOPIC_FOR_DELETE = new Topic().setName("topic-to-delete-" + randomAlphabetic(5)); + private static final List<Topic> TOPIC_LIST = new ArrayList<>(); + + @BeforeAll + public void beforeAll() { + TOPIC_LIST.addAll(List.of(TOPIC_FOR_UPDATE, TOPIC_FOR_DELETE)); + TOPIC_LIST.forEach(topic -> apiHelper.createTopic(CLUSTER_NAME, topic.getName())); + } + + @DisplayName("should create a topic") + @Suite(suiteId = 4, title = "Create new Topic") + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(199) + @Test + public void createTopic() { + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .clickAddTopicBtn(); + topicCreateEditForm + .waitUntilScreenReady() + .setTopicName(TOPIC_TO_CREATE.getName()) + .setPartitions(TOPIC_TO_CREATE.getPartitions()) + .selectCleanupPolicy(TOPIC_TO_CREATE.getCleanupPolicyValue()) + .clickCreateTopicBtn(); + topicDetails + .waitUntilScreenReady(); + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .openTopic(TOPIC_TO_CREATE.getName()); + SoftAssertions softly = new SoftAssertions(); + softly.assertThat(topicDetails.isTopicHeaderVisible(TOPIC_TO_CREATE.getName())).as("isTopicHeaderVisible()").isTrue(); + softly.assertThat(topicDetails.getCleanUpPolicy()).as("getCleanUpPolicy()").isEqualTo(TOPIC_TO_CREATE.getCleanupPolicyValue().toString()); + softly.assertThat(topicDetails.getPartitions()).as("getPartitions()").isEqualTo(TOPIC_TO_CREATE.getPartitions()); + softly.assertAll(); + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady(); + Assertions.assertTrue(topicsList.isTopicVisible(TOPIC_TO_CREATE.getName()), "isTopicVisible"); + TOPIC_LIST.add(TOPIC_TO_CREATE); + } + + @Disabled() + @Issue("https://github.com/provectus/kafka-ui/issues/2625") + @DisplayName("should update a topic") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(197) + @Test + public void updateTopic() { + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .openTopic(TOPIC_FOR_UPDATE.getName()); + topicDetails + .waitUntilScreenReady() + .openDotMenu() + .clickEditSettingsMenu(); + topicCreateEditForm + .waitUntilScreenReady() + .selectCleanupPolicy((TOPIC_FOR_UPDATE.getCleanupPolicyValue())) + .setMinInsyncReplicas(10) + .setTimeToRetainDataInMs(TOPIC_FOR_UPDATE.getTimeToRetainData()) + .setMaxSizeOnDiskInGB(TOPIC_FOR_UPDATE.getMaxSizeOnDisk()) + .setMaxMessageBytes(TOPIC_FOR_UPDATE.getMaxMessageBytes()) + .clickCreateTopicBtn(); + topicDetails + .waitUntilScreenReady(); + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .openTopic(TOPIC_FOR_UPDATE.getName()); + topicDetails + .waitUntilScreenReady() + .openDotMenu() + .clickEditSettingsMenu(); + SoftAssertions softly = new SoftAssertions(); + softly.assertThat(topicCreateEditForm.getCleanupPolicy()).as("getCleanupPolicy()").isEqualTo(TOPIC_FOR_UPDATE.getCleanupPolicyValue().getVisibleText()); + softly.assertThat(topicCreateEditForm.getTimeToRetain()).as("getTimeToRetain()").isEqualTo(TOPIC_FOR_UPDATE.getTimeToRetainData()); + softly.assertThat(topicCreateEditForm.getMaxSizeOnDisk()).as("getMaxSizeOnDisk()").isEqualTo(TOPIC_FOR_UPDATE.getMaxSizeOnDisk().getVisibleText()); + softly.assertThat(topicCreateEditForm.getMaxMessageBytes()).as("getMaxMessageBytes()").isEqualTo(TOPIC_FOR_UPDATE.getMaxMessageBytes()); + softly.assertAll(); + } + + @DisplayName("should delete topic") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(207) + @Test + public void deleteTopic() { + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .openTopic(TOPIC_FOR_DELETE.getName()); + topicDetails + .waitUntilScreenReady() + .openDotMenu() + .clickDeleteTopicMenu() + .clickConfirmDeleteBtn(); + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady(); + Assertions.assertFalse(topicsList.isTopicVisible(TOPIC_FOR_DELETE.getName()), "isTopicVisible"); + TOPIC_LIST.remove(TOPIC_FOR_DELETE); + } + + @DisplayName("Redirect to consumer from topic profile") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(20) + @Test + void redirectToConsumerFromTopic() { + String topicName = "source-activities"; + String consumerGroupId = "connect-sink_postgres_activities"; + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .openTopic(topicName); + topicDetails + .waitUntilScreenReady() + .openDetailsTab(TopicDetails.TopicMenu.CONSUMERS) + .openConsumerGroup(consumerGroupId); + consumersDetails + .waitUntilScreenReady(); + assertThat(consumersDetails.isRedirectedConsumerTitleVisible(consumerGroupId)) + .withFailMessage("isRedirectedConsumerTitleVisible").isTrue(); + assertThat(consumersDetails.isTopicInConsumersDetailsVisible(topicName)) + .withFailMessage("isTopicInConsumersDetailsVisible").isTrue(); + } + + @DisplayName("Checking Topic creation possibility in case of empty Topic Name") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(4) + @Test + void checkTopicCreatePossibility() { + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .clickAddTopicBtn(); + topicCreateEditForm + .waitUntilScreenReady() + .setTopicName(""); + assertThat(topicCreateEditForm.isCreateTopicButtonEnabled()).as("isCreateTopicButtonEnabled()").isFalse(); + topicCreateEditForm + .setTopicName("testTopic1"); + assertThat(topicCreateEditForm.isCreateTopicButtonEnabled()).as("isCreateTopicButtonEnabled()").isTrue(); + } + + @DisplayName("Checking requiredness of Custom parameters within 'Create new Topic'") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(6) + @Test + void checkCustomParametersWithinCreateNewTopic() { + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .clickAddTopicBtn(); + topicCreateEditForm + .waitUntilScreenReady() + .setTopicName(TOPIC_TO_CREATE.getName()) + .clickAddCustomParameterTypeButton() + .setCustomParameterType(TOPIC_TO_CREATE.getCustomParameterType()); + assertThat(topicCreateEditForm.isDeleteCustomParameterButtonEnabled()).as("isDeleteCustomParameterButtonEnabled()") + .isTrue(); + topicCreateEditForm + .clearCustomParameterValue(); + assertThat(topicCreateEditForm.isValidationMessageCustomParameterValueVisible()) + .as("isValidationMessageCustomParameterValueVisible()").isTrue(); + } + + @DisplayName("Checking Topics section within Kafka-ui Application") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(2) + @Test + void checkTopicListElements(){ + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady(); + SoftAssertions softly = new SoftAssertions(); + topicsList.getAllVisibleElements().forEach( + element -> softly.assertThat(element.is(Condition.visible)).as(element.getSearchCriteria() + " isVisible()") + .isTrue()); + topicsList.getAllEnabledElements().forEach( + element -> softly.assertThat(element.is(Condition.enabled)).as(element.getSearchCriteria() + " isEnabled()") + .isTrue()); + softly.assertAll(); + } + + @AfterAll + public void afterAll() { + TOPIC_LIST.forEach(topic -> apiHelper.deleteTopic(CLUSTER_NAME, topic.getName())); + } +} diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicMessagesTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicMessagesTests.java new file mode 100644 index 00000000000..f3da20747be --- /dev/null +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicMessagesTests.java @@ -0,0 +1,133 @@ +package com.provectus.kafka.ui.suite.topics; + +import static com.provectus.kafka.ui.pages.BasePage.AlertHeader.SUCCESS; +import static com.provectus.kafka.ui.pages.NaviSideBar.SideMenuOption.TOPICS; +import static com.provectus.kafka.ui.settings.Source.CLUSTER_NAME; +import static com.provectus.kafka.ui.utilities.FileUtils.fileToString; +import static org.apache.commons.lang.RandomStringUtils.randomAlphabetic; + +import com.provectus.kafka.ui.base.BaseTest; +import com.provectus.kafka.ui.models.Topic; +import com.provectus.kafka.ui.pages.topic.TopicDetails; +import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.AutomationStatus; +import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.Suite; +import com.provectus.kafka.ui.utilities.qaseIoUtils.enums.Status; +import io.qameta.allure.Issue; +import io.qase.api.annotation.CaseId; +import java.util.ArrayList; +import java.util.List; +import org.assertj.core.api.SoftAssertions; +import org.junit.jupiter.api.AfterAll; +import org.junit.jupiter.api.Assertions; +import org.junit.jupiter.api.BeforeAll; +import org.junit.jupiter.api.Disabled; +import org.junit.jupiter.api.DisplayName; +import org.junit.jupiter.api.Test; +import org.junit.jupiter.api.TestInstance; + +@TestInstance(TestInstance.Lifecycle.PER_CLASS) +public class TopicMessagesTests extends BaseTest { + private static final long SUITE_ID = 2; + private static final String SUITE_TITLE = "Topics"; + private static final Topic TOPIC_FOR_MESSAGES = new Topic() + .setName("topic-with-clean-message-attribute-" + randomAlphabetic(5)) + .setMessageKey(fileToString(System.getProperty("user.dir") + "/src/test/resources/producedkey.txt")) + .setMessageContent(fileToString(System.getProperty("user.dir") + "/src/test/resources/testData.txt")); + private static final List<Topic> TOPIC_LIST = new ArrayList<>(); + + @BeforeAll + public void beforeAll() { + TOPIC_LIST.addAll(List.of(TOPIC_FOR_MESSAGES)); + TOPIC_LIST.forEach(topic -> apiHelper.createTopic(CLUSTER_NAME, topic.getName())); + } + + @DisplayName("produce message") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(222) + @Test + void produceMessage() { + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .openTopic(TOPIC_FOR_MESSAGES.getName()); + topicDetails + .waitUntilScreenReady() + .openDetailsTab(TopicDetails.TopicMenu.MESSAGES) + .clickProduceMessageBtn(); + produceMessagePanel + .waitUntilScreenReady() + .setContentFiled(TOPIC_FOR_MESSAGES.getMessageContent()) + .setKeyField(TOPIC_FOR_MESSAGES.getMessageKey()) + .submitProduceMessage(); + topicDetails + .waitUntilScreenReady(); + SoftAssertions softly = new SoftAssertions(); + softly.assertThat(topicDetails.isKeyMessageVisible((TOPIC_FOR_MESSAGES.getMessageKey()))).withFailMessage("isKeyMessageVisible()").isTrue(); + softly.assertThat(topicDetails.isContentMessageVisible((TOPIC_FOR_MESSAGES.getMessageContent()).trim())).withFailMessage("isContentMessageVisible()").isTrue(); + softly.assertAll(); + } + + @Disabled + @Issue("https://github.com/provectus/kafka-ui/issues/2778") + @DisplayName("clear message") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(19) + @Test + void clearMessage() { + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .openTopic(TOPIC_FOR_MESSAGES.getName()); + topicDetails + .waitUntilScreenReady() + .openDetailsTab(TopicDetails.TopicMenu.OVERVIEW) + .clickProduceMessageBtn(); + int messageAmount = topicDetails.getMessageCountAmount(); + produceMessagePanel + .waitUntilScreenReady() + .setContentFiled(TOPIC_FOR_MESSAGES.getMessageContent()) + .setKeyField(TOPIC_FOR_MESSAGES.getMessageKey()) + .submitProduceMessage(); + topicDetails + .waitUntilScreenReady(); + Assertions.assertEquals(messageAmount + 1, topicDetails.getMessageCountAmount(), "getMessageCountAmount()"); + topicDetails + .openDotMenu() + .clickClearMessagesMenu() + .waitUntilScreenReady(); + Assertions.assertEquals(0, topicDetails.getMessageCountAmount(), "getMessageCountAmount()"); + } + + @Disabled + @Issue("https://github.com/provectus/kafka-ui/issues/2819") + @DisplayName("Message copy from topic profile") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(21) + @Test + void copyMessageFromTopicProfile() { + String topicName = "_schemas"; + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .openTopic(topicName); + topicDetails + .waitUntilScreenReady() + .openDetailsTab(TopicDetails.TopicMenu.MESSAGES) + .getRandomMessage() + .openDotMenu() + .clickCopyToClipBoard(); + Assertions.assertTrue(topicDetails.isAlertWithMessageVisible(SUCCESS, "Copied successfully!"), + "isAlertWithMessageVisible()"); + } + + @AfterAll + public void afterAll() { + TOPIC_LIST.forEach(topic -> apiHelper.deleteTopic(CLUSTER_NAME, topic.getName())); + } +} diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java deleted file mode 100644 index dc7a3b130a8..00000000000 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/tests/TopicTests.java +++ /dev/null @@ -1,329 +0,0 @@ -package com.provectus.kafka.ui.tests; - -import static com.provectus.kafka.ui.pages.BasePage.AlertHeader.SUCCESS; -import static com.provectus.kafka.ui.pages.NaviSideBar.SideMenuOption.TOPICS; -import static com.provectus.kafka.ui.pages.topic.enums.CleanupPolicyValue.COMPACT; -import static com.provectus.kafka.ui.pages.topic.enums.CleanupPolicyValue.DELETE; -import static com.provectus.kafka.ui.pages.topic.enums.CustomParameterType.COMPRESSION_TYPE; -import static com.provectus.kafka.ui.pages.topic.enums.MaxSizeOnDisk.SIZE_20_GB; -import static com.provectus.kafka.ui.settings.Source.CLUSTER_NAME; -import static com.provectus.kafka.ui.utilities.FileUtils.fileToString; -import static org.apache.commons.lang.RandomStringUtils.randomAlphabetic; -import static org.assertj.core.api.Assertions.assertThat; - -import com.provectus.kafka.ui.base.BaseTest; -import com.provectus.kafka.ui.models.Topic; -import com.provectus.kafka.ui.pages.BasePage; -import com.provectus.kafka.ui.pages.topic.TopicDetails; -import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.AutomationStatus; -import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.Suite; -import com.provectus.kafka.ui.utilities.qaseIoUtils.enums.Status; -import io.qameta.allure.Issue; -import io.qase.api.annotation.CaseId; -import java.util.ArrayList; -import java.util.List; -import org.assertj.core.api.SoftAssertions; -import org.junit.jupiter.api.AfterAll; -import org.junit.jupiter.api.Assertions; -import org.junit.jupiter.api.BeforeAll; -import org.junit.jupiter.api.Disabled; -import org.junit.jupiter.api.DisplayName; -import org.junit.jupiter.api.Test; -import org.junit.jupiter.api.TestInstance; - -@TestInstance(TestInstance.Lifecycle.PER_CLASS) -public class TopicTests extends BaseTest { - private static final long SUITE_ID = 2; - private static final String SUITE_TITLE = "Topics"; - private static final Topic TOPIC_TO_CREATE = new Topic() - .setName("new-topic-"+ randomAlphabetic(5)) - .setPartitions("1") - .setCustomParameterType(COMPRESSION_TYPE) - .setCustomParameterValue("producer") - .setCleanupPolicyValue(DELETE); - private static final Topic TOPIC_FOR_UPDATE = new Topic() - .setName("topic-to-update-" + randomAlphabetic(5)) - .setCleanupPolicyValue(COMPACT) - .setTimeToRetainData("604800001") - .setMaxSizeOnDisk(SIZE_20_GB) - .setMaxMessageBytes("1000020") - .setMessageKey(fileToString(System.getProperty("user.dir") + "/src/test/resources/producedkey.txt")) - .setMessageContent(fileToString(System.getProperty("user.dir") + "/src/test/resources/testData.txt")); - private static final Topic TOPIC_FOR_MESSAGES = new Topic() - .setName("topic-with-clean-message-attribute-" + randomAlphabetic(5)) - .setMessageKey(fileToString(System.getProperty("user.dir") + "/src/test/resources/producedkey.txt")) - .setMessageContent(fileToString(System.getProperty("user.dir") + "/src/test/resources/testData.txt")); - - private static final Topic TOPIC_FOR_DELETE = new Topic().setName("topic-to-delete-" + randomAlphabetic(5)); - private static final List<Topic> TOPIC_LIST = new ArrayList<>(); - - @BeforeAll - public void beforeAll() { - TOPIC_LIST.addAll(List.of(TOPIC_FOR_UPDATE, TOPIC_FOR_DELETE, TOPIC_FOR_MESSAGES)); - TOPIC_LIST.forEach(topic -> apiHelper.createTopic(CLUSTER_NAME, topic.getName())); - } - - @DisplayName("should create a topic") - @Suite(suiteId = 4, title = "Create new Topic") - @AutomationStatus(status = Status.AUTOMATED) - @CaseId(199) - @Test - public void createTopic() { - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady() - .clickAddTopicBtn(); - topicCreateEditForm - .waitUntilScreenReady() - .setTopicName(TOPIC_TO_CREATE.getName()) - .setPartitions(TOPIC_TO_CREATE.getPartitions()) - .selectCleanupPolicy(TOPIC_TO_CREATE.getCleanupPolicyValue()) - .clickCreateTopicBtn(); - topicDetails - .waitUntilScreenReady(); - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady() - .openTopic(TOPIC_TO_CREATE.getName()); - SoftAssertions softly = new SoftAssertions(); - softly.assertThat(topicDetails.isTopicHeaderVisible(TOPIC_TO_CREATE.getName())).as("isTopicHeaderVisible()").isTrue(); - softly.assertThat(topicDetails.getCleanUpPolicy()).as("getCleanUpPolicy()").isEqualTo(TOPIC_TO_CREATE.getCleanupPolicyValue().toString()); - softly.assertThat(topicDetails.getPartitions()).as("getPartitions()").isEqualTo(TOPIC_TO_CREATE.getPartitions()); - softly.assertAll(); - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady(); - Assertions.assertTrue(topicsList.isTopicVisible(TOPIC_TO_CREATE.getName()), "isTopicVisible"); - TOPIC_LIST.add(TOPIC_TO_CREATE); - } - - @Disabled() - @Issue("https://github.com/provectus/kafka-ui/issues/2625") - @DisplayName("should update a topic") - @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) - @AutomationStatus(status = Status.AUTOMATED) - @CaseId(197) - @Test - public void updateTopic() { - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady() - .openTopic(TOPIC_FOR_UPDATE.getName()); - topicDetails - .waitUntilScreenReady() - .openDotMenu() - .clickEditSettingsMenu(); - topicCreateEditForm - .waitUntilScreenReady() - .selectCleanupPolicy((TOPIC_FOR_UPDATE.getCleanupPolicyValue())) - .setMinInsyncReplicas(10) - .setTimeToRetainDataInMs(TOPIC_FOR_UPDATE.getTimeToRetainData()) - .setMaxSizeOnDiskInGB(TOPIC_FOR_UPDATE.getMaxSizeOnDisk()) - .setMaxMessageBytes(TOPIC_FOR_UPDATE.getMaxMessageBytes()) - .clickCreateTopicBtn(); - topicDetails - .waitUntilScreenReady(); - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady() - .openTopic(TOPIC_FOR_UPDATE.getName()); - topicDetails - .waitUntilScreenReady() - .openDotMenu() - .clickEditSettingsMenu(); - SoftAssertions softly = new SoftAssertions(); - softly.assertThat(topicCreateEditForm.getCleanupPolicy()).as("getCleanupPolicy()").isEqualTo(TOPIC_FOR_UPDATE.getCleanupPolicyValue().getVisibleText()); - softly.assertThat(topicCreateEditForm.getTimeToRetain()).as("getTimeToRetain()").isEqualTo(TOPIC_FOR_UPDATE.getTimeToRetainData()); - softly.assertThat(topicCreateEditForm.getMaxSizeOnDisk()).as("getMaxSizeOnDisk()").isEqualTo(TOPIC_FOR_UPDATE.getMaxSizeOnDisk().getVisibleText()); - softly.assertThat(topicCreateEditForm.getMaxMessageBytes()).as("getMaxMessageBytes()").isEqualTo(TOPIC_FOR_UPDATE.getMaxMessageBytes()); - softly.assertAll(); - } - - @DisplayName("should delete topic") - @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) - @AutomationStatus(status = Status.AUTOMATED) - @CaseId(207) - @Test - public void deleteTopic() { - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady() - .openTopic(TOPIC_FOR_DELETE.getName()); - topicDetails - .waitUntilScreenReady() - .openDotMenu() - .clickDeleteTopicMenu() - .clickConfirmDeleteBtn(); - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady(); - Assertions.assertFalse(topicsList.isTopicVisible(TOPIC_FOR_DELETE.getName()), "isTopicVisible"); - TOPIC_LIST.remove(TOPIC_FOR_DELETE); - } - - @DisplayName("produce message") - @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) - @AutomationStatus(status = Status.AUTOMATED) - @CaseId(222) - @Test - void produceMessage() { - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady() - .openTopic(TOPIC_FOR_MESSAGES.getName()); - topicDetails - .waitUntilScreenReady() - .openDetailsTab(TopicDetails.TopicMenu.MESSAGES) - .clickProduceMessageBtn(); - produceMessagePanel - .waitUntilScreenReady() - .setContentFiled(TOPIC_FOR_MESSAGES.getMessageContent()) - .setKeyField(TOPIC_FOR_MESSAGES.getMessageKey()) - .submitProduceMessage(); - topicDetails - .waitUntilScreenReady(); - SoftAssertions softly = new SoftAssertions(); - softly.assertThat(topicDetails.isKeyMessageVisible((TOPIC_FOR_MESSAGES.getMessageKey()))).withFailMessage("isKeyMessageVisible()").isTrue(); - softly.assertThat(topicDetails.isContentMessageVisible((TOPIC_FOR_MESSAGES.getMessageContent()).trim())).withFailMessage("isContentMessageVisible()").isTrue(); - softly.assertAll(); - } - - @Disabled - @Issue("https://github.com/provectus/kafka-ui/issues/2778") - @DisplayName("clear message") - @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) - @AutomationStatus(status = Status.AUTOMATED) - @CaseId(19) - @Test - void clearMessage() { - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady() - .openTopic(TOPIC_FOR_MESSAGES.getName()); - topicDetails - .waitUntilScreenReady() - .openDetailsTab(TopicDetails.TopicMenu.OVERVIEW) - .clickProduceMessageBtn(); - int messageAmount = topicDetails.getMessageCountAmount(); - produceMessagePanel - .waitUntilScreenReady() - .setContentFiled(TOPIC_FOR_MESSAGES.getMessageContent()) - .setKeyField(TOPIC_FOR_MESSAGES.getMessageKey()) - .submitProduceMessage(); - topicDetails - .waitUntilScreenReady(); - Assertions.assertEquals(messageAmount + 1, topicDetails.getMessageCountAmount(), "getMessageCountAmount()"); - topicDetails - .openDotMenu() - .clickClearMessagesMenu() - .waitUntilScreenReady(); - Assertions.assertEquals(0, topicDetails.getMessageCountAmount(), "getMessageCountAmount()"); - } - - @DisplayName("Redirect to consumer from topic profile") - @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) - @AutomationStatus(status = Status.AUTOMATED) - @CaseId(20) - @Test - void redirectToConsumerFromTopic() { - String topicName = "source-activities"; - String consumerGroupId = "connect-sink_postgres_activities"; - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady() - .openTopic(topicName); - topicDetails - .waitUntilScreenReady() - .openDetailsTab(TopicDetails.TopicMenu.CONSUMERS) - .openConsumerGroup(consumerGroupId); - consumersDetails - .waitUntilScreenReady(); - assertThat(consumersDetails.isRedirectedConsumerTitleVisible(consumerGroupId)) - .withFailMessage("isRedirectedConsumerTitleVisible").isTrue(); - assertThat(consumersDetails.isTopicInConsumersDetailsVisible(topicName)) - .withFailMessage("isTopicInConsumersDetailsVisible").isTrue(); - } - - @DisplayName("Checking Topic creation possibility in case of empty Topic Name") - @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) - @AutomationStatus(status = Status.AUTOMATED) - @CaseId(4) - @Test - void checkTopicCreatePossibility() { - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady() - .clickAddTopicBtn(); - topicCreateEditForm - .waitUntilScreenReady() - .setTopicName(""); - assertThat(topicCreateEditForm.isCreateTopicButtonEnabled()).as("isCreateTopicButtonEnabled()").isFalse(); - topicCreateEditForm - .setTopicName("testTopic1"); - assertThat(topicCreateEditForm.isCreateTopicButtonEnabled()).as("isCreateTopicButtonEnabled()").isTrue(); - } - - @DisplayName("Checking requiredness of Custom parameters within 'Create new Topic'") - @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) - @AutomationStatus(status = Status.AUTOMATED) - @CaseId(6) - @Test - void checkCustomParametersWithinCreateNewTopic() { - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady() - .clickAddTopicBtn(); - topicCreateEditForm - .waitUntilScreenReady() - .setTopicName(TOPIC_TO_CREATE.getName()) - .clickAddCustomParameterTypeButton() - .setCustomParameterType(TOPIC_TO_CREATE.getCustomParameterType()); - assertThat(topicCreateEditForm.isDeleteCustomParameterButtonEnabled()).as("isDeleteCustomParameterButtonEnabled()") - .isTrue(); - topicCreateEditForm - .clearCustomParameterValue(); - assertThat(topicCreateEditForm.isValidationMessageCustomParameterValueVisible()) - .as("isValidationMessageCustomParameterValueVisible()").isTrue(); - } - - @Disabled - @Issue("https://github.com/provectus/kafka-ui/issues/2819") - @DisplayName("Message copy from topic profile") - @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) - @AutomationStatus(status = Status.AUTOMATED) - @CaseId(21) - @Test - void copyMessageFromTopicProfile() { - String topicName = "_schemas"; - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady() - .openTopic(topicName); - topicDetails - .waitUntilScreenReady() - .openDetailsTab(TopicDetails.TopicMenu.MESSAGES) - .getRandomMessage() - .openDotMenu() - .clickCopyToClipBoard(); - Assertions.assertTrue(topicDetails.isAlertWithMessageVisible(SUCCESS, "Copied successfully!"), - "isAlertWithMessageVisible()"); - } - - @AfterAll - public void afterAll() { - TOPIC_LIST.forEach(topic -> apiHelper.deleteTopic(CLUSTER_NAME, topic.getName())); - } -}
train
val
2022-11-11T12:43:15
"2022-11-09T13:42:46Z"
anezboretskiy
train
provectus/kafka-ui/2711_2926
provectus/kafka-ui
provectus/kafka-ui/2711
provectus/kafka-ui/2926
[ "keyword_issue_to_pr", "connected" ]
6de731778bf9e2884f5f50a464d4f399055de81e
f47a0382f73fc88b778d87492d09bc9b8427da1c
[ "Can I work on this and raise a PR ?", "Hey, sure.", "@royca If you don't mind I would like to work on this since there has been no PR or any discussion for 12 days\r\n@Haarolean please assign to me this issue if it possible", "A fix for\r\nCVE-2022-31690\r\nCVE-2022-31692\r\nNew dependency [unavailable yet](https://github.com/spring-projects/spring-security/issues/12131), waiting for a spring boot 2.7.6.\r\n\r\nFixes for:\r\nCVE-2022-3171\r\nCVE-2022-36944\r\nGHSA-h4h5-3hr4-j3g2\r\nare unavailable yet, blame/wait for confluent\r\n\r\nSnake yaml CVEs:\r\nCVE-2022-25857 \r\nCVE-2022-38749 \r\nCVE-2022-38750 \r\nCVE-2022-38751 \r\n CVE-2022-38752 \r\n\r\nare [false positives](https://github.com/spring-projects/spring-boot/issues/32221).\r\n\r\nOthers have been fixed within #2926", "@Haarolean I could see snakeyaml CVEs are false positive. Is it? I could see the issues created \r\n\r\nHere is an example \r\nhttps://bitbucket.org/snakeyaml/snakeyaml/issues/525 and they have been resolved.\r\n\r\nSo, is it possible to provide snakeyaml version 1.33 in 2.7.x version?", "@Subrhamanya yes they are according to [this thread](https://github.com/spring-projects/spring-boot/issues/32221).\r\nAnyway, we've bumped some dependencies once again and now we have a newer transitive snakeyaml version.", "Does that mean 2.7.6 version of spring-boot will have snakeyaml version 1.33?", "> Does that mean 2.7.6 version of spring-boot will have snakeyaml version 1.33?\r\n\r\nAFAIK, yes." ]
[]
"2022-11-11T16:48:43Z"
[ "good first issue", "scope/backend", "type/security", "status/accepted" ]
CVEs fixes, October 22
https://github.com/provectus/kafka-ui/actions/runs/3058781951/jobs/4935377778 Bump snakeyaml deps
[ "pom.xml" ]
[ "pom.xml" ]
[]
diff --git a/pom.xml b/pom.xml index 73d3bfeb3ad..ecb707e1da0 100644 --- a/pom.xml +++ b/pom.xml @@ -14,7 +14,7 @@ <maven.compiler.release>17</maven.compiler.release> <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding> - <spring-boot.version>2.6.8</spring-boot.version> + <spring-boot.version>2.7.5</spring-boot.version> <jackson-databind-nullable.version>0.2.2</jackson-databind-nullable.version> <org.mapstruct.version>1.4.2.Final</org.mapstruct.version> <org.projectlombok.version>1.18.20</org.projectlombok.version> @@ -33,7 +33,7 @@ <swagger-annotations.version>1.6.0</swagger-annotations.version> <springdoc-openapi-webflux-ui.version>1.2.32</springdoc-openapi-webflux-ui.version> <avro.version>1.11.0</avro.version> - <confluent.version>7.0.1</confluent.version> + <confluent.version>7.3.0</confluent.version> <apache.commons.version>2.11.1</apache.commons.version> <test.containers.version>1.17.1</test.containers.version> <okhttp3.mockwebserver.version>4.10.0</okhttp3.mockwebserver.version>
null
train
val
2022-11-11T14:19:11
"2022-10-07T08:47:40Z"
Haarolean
train
provectus/kafka-ui/2711_2929
provectus/kafka-ui
provectus/kafka-ui/2711
provectus/kafka-ui/2929
[ "keyword_pr_to_issue", "connected" ]
818149977cc234bf18ba9eeabb00926c16610bc4
0d0a24491586a499890897dac5f3546510f53cd5
[ "Can I work on this and raise a PR ?", "Hey, sure.", "@royca If you don't mind I would like to work on this since there has been no PR or any discussion for 12 days\r\n@Haarolean please assign to me this issue if it possible", "A fix for\r\nCVE-2022-31690\r\nCVE-2022-31692\r\nNew dependency [unavailable yet](https://github.com/spring-projects/spring-security/issues/12131), waiting for a spring boot 2.7.6.\r\n\r\nFixes for:\r\nCVE-2022-3171\r\nCVE-2022-36944\r\nGHSA-h4h5-3hr4-j3g2\r\nare unavailable yet, blame/wait for confluent\r\n\r\nSnake yaml CVEs:\r\nCVE-2022-25857 \r\nCVE-2022-38749 \r\nCVE-2022-38750 \r\nCVE-2022-38751 \r\n CVE-2022-38752 \r\n\r\nare [false positives](https://github.com/spring-projects/spring-boot/issues/32221).\r\n\r\nOthers have been fixed within #2926", "@Haarolean I could see snakeyaml CVEs are false positive. Is it? I could see the issues created \r\n\r\nHere is an example \r\nhttps://bitbucket.org/snakeyaml/snakeyaml/issues/525 and they have been resolved.\r\n\r\nSo, is it possible to provide snakeyaml version 1.33 in 2.7.x version?", "@Subrhamanya yes they are according to [this thread](https://github.com/spring-projects/spring-boot/issues/32221).\r\nAnyway, we've bumped some dependencies once again and now we have a newer transitive snakeyaml version.", "Does that mean 2.7.6 version of spring-boot will have snakeyaml version 1.33?", "> Does that mean 2.7.6 version of spring-boot will have snakeyaml version 1.33?\r\n\r\nAFAIK, yes." ]
[ "any specific reason we're downgrading?", "@Haarolean Outdated local `master` branch. 🀦 \r\n\r\nI've updated the PR.", "@joschi any idea how did this thing get in here? There's no scala library prior to your commit πŸ€” ", "@Haarolean Kafka (`org.apache.kafka:kafka_2.13` used in `kafka-ui-e2e-checks`) is written in Scala and was pulling in a vulnerable version.\r\n\r\nThis property is only used in dependency management, so that Maven will pull in the non-vulnerable version of the Scala standard library instead.\r\n\r\nhttps://github.com/personio/kafka-ui/blob/2104d81e139bb6b489594dc10caf676ea725dcf9/pom.xml#L132-L136", "@joschi thank you very much, I prefer to indicate this in poms, I'll add the comment" ]
"2022-11-14T09:27:07Z"
[ "good first issue", "scope/backend", "type/security", "status/accepted" ]
CVEs fixes, October 22
https://github.com/provectus/kafka-ui/actions/runs/3058781951/jobs/4935377778 Bump snakeyaml deps
[ "kafka-ui-api/pom.xml", "kafka-ui-contract/pom.xml", "kafka-ui-e2e-checks/pom.xml", "kafka-ui-serde-api/pom.xml", "pom.xml" ]
[ "kafka-ui-api/pom.xml", "kafka-ui-contract/pom.xml", "kafka-ui-e2e-checks/pom.xml", "kafka-ui-serde-api/pom.xml", "pom.xml" ]
[]
diff --git a/kafka-ui-api/pom.xml b/kafka-ui-api/pom.xml index f9fa40386fb..95c64071dc5 100644 --- a/kafka-ui-api/pom.xml +++ b/kafka-ui-api/pom.xml @@ -20,18 +20,6 @@ <sonar.language>java</sonar.language> </properties> - <dependencyManagement> - <dependencies> - <dependency> - <groupId>org.springframework.boot</groupId> - <artifactId>spring-boot-dependencies</artifactId> - <version>${spring-boot.version}</version> - <type>pom</type> - <scope>import</scope> - </dependency> - </dependencies> - </dependencyManagement> - <dependencies> <dependency> <groupId>org.springframework.boot</groupId> @@ -103,7 +91,7 @@ <dependency> <groupId>software.amazon.msk</groupId> <artifactId>aws-msk-iam-auth</artifactId> - <version>1.1.3</version> + <version>1.1.5</version> </dependency> <dependency> @@ -140,25 +128,21 @@ <dependency> <groupId>org.testcontainers</groupId> <artifactId>testcontainers</artifactId> - <version>${test.containers.version}</version> <scope>test</scope> </dependency> <dependency> <groupId>org.testcontainers</groupId> <artifactId>kafka</artifactId> - <version>${test.containers.version}</version> <scope>test</scope> </dependency> <dependency> <groupId>org.testcontainers</groupId> <artifactId>junit-jupiter</artifactId> - <version>${test.containers.version}</version> <scope>test</scope> </dependency> <dependency> <groupId>org.junit.jupiter</groupId> <artifactId>junit-jupiter-engine</artifactId> - <version>${junit-jupiter-engine.version}</version> <scope>test</scope> </dependency> <dependency> diff --git a/kafka-ui-contract/pom.xml b/kafka-ui-contract/pom.xml index 7f0863ed3b5..3c898b4fcf4 100644 --- a/kafka-ui-contract/pom.xml +++ b/kafka-ui-contract/pom.xml @@ -21,12 +21,10 @@ <dependency> <groupId>org.springframework.boot</groupId> <artifactId>spring-boot-starter-webflux</artifactId> - <version>${spring-boot.version}</version> </dependency> <dependency> <groupId>org.springframework.boot</groupId> <artifactId>spring-boot-starter-validation</artifactId> - <version>${spring-boot.version}</version> </dependency> <dependency> <groupId>io.swagger</groupId> diff --git a/kafka-ui-e2e-checks/pom.xml b/kafka-ui-e2e-checks/pom.xml index 54a41c56ca3..12a702a76a5 100644 --- a/kafka-ui-e2e-checks/pom.xml +++ b/kafka-ui-e2e-checks/pom.xml @@ -12,27 +12,22 @@ <artifactId>kafka-ui-e2e-checks</artifactId> <properties> <kafka-ui-contract>${project.version}</kafka-ui-contract> - <junit.version>5.8.2</junit.version> <aspectj.version>1.9.8</aspectj.version> <allure.version>2.18.1</allure.version> <json-smart.version>2.4.8</json-smart.version> - <testcontainers.version>1.17.2</testcontainers.version> <selenide.version>6.6.3</selenide.version> <assertj.version>3.23.1</assertj.version> <google.auto-service.version>1.0.1</google.auto-service.version> <hamcrest.version>2.2</hamcrest.version> <slf4j.version>1.7.36</slf4j.version> - <testcontainers.junit-jupiter.version>1.17.2</testcontainers.junit-jupiter.version> <allure.java-commons.version>2.18.1</allure.java-commons.version> <dotenv.version>2.2.4</dotenv.version> - <junit.platform-launcher.version>1.8.2</junit.platform-launcher.version> <allure.maven-plugin.version>2.6</allure.maven-plugin.version> <ashot.version>1.5.4</ashot.version> <allure.screendiff-plugin.version>2.18.1</allure.screendiff-plugin.version> <maven.surefire-plugin.version>2.22.2</maven.surefire-plugin.version> <allure-maven.version>2.10.0</allure-maven.version> - <kafka.version>3.0.0</kafka.version> - <netty.version>4.1.84.Final</netty.version> + <kafka.version>3.3.1</kafka.version> <qase.io.version>2.1.3</qase.io.version> </properties> @@ -89,54 +84,44 @@ <dependency> <groupId>io.netty</groupId> <artifactId>netty-buffer</artifactId> - <version>${netty.version}</version> </dependency> <dependency> <groupId>io.netty</groupId> <artifactId>netty-common</artifactId> - <version>${netty.version}</version> </dependency> <dependency> <groupId>io.netty</groupId> <artifactId>netty-codec</artifactId> - <version>${netty.version}</version> </dependency> <dependency> <groupId>io.netty</groupId> <artifactId>netty-handler</artifactId> - <version>${netty.version}</version> </dependency> <dependency> <groupId>io.netty</groupId> <artifactId>netty-resolver</artifactId> - <version>${netty.version}</version> </dependency> <dependency> <groupId>io.netty</groupId> <artifactId>netty-transport</artifactId> - <version>${netty.version}</version> </dependency> <dependency> <groupId>io.netty</groupId> <artifactId>netty-transport-native-epoll</artifactId> - <version>${netty.version}</version> </dependency> <dependency> <groupId>io.netty</groupId> <artifactId>netty-transport-native-unix-common</artifactId> - <version>${netty.version}</version> </dependency> <dependency> <groupId>io.netty</groupId> <artifactId>netty-resolver-dns-native-macos</artifactId> - <version>${netty.version}</version> <classifier>osx-aarch_64</classifier> </dependency> <dependency> <groupId>org.testcontainers</groupId> <artifactId>testcontainers</artifactId> - <version>${testcontainers.version}</version> </dependency> <dependency> @@ -199,7 +184,6 @@ <dependency> <groupId>org.testcontainers</groupId> <artifactId>junit-jupiter</artifactId> - <version>${testcontainers.junit-jupiter.version}</version> </dependency> <dependency> <groupId>io.qameta.allure</groupId> @@ -214,7 +198,6 @@ <dependency> <groupId>org.junit.platform</groupId> <artifactId>junit-platform-launcher</artifactId> - <version>${junit.platform-launcher.version}</version> </dependency> <dependency> <groupId>ru.yandex.qatools.allure</groupId> @@ -245,7 +228,6 @@ <dependency> <groupId>org.testcontainers</groupId> <artifactId>selenium</artifactId> - <version>${testcontainers.version}</version> <scope>test</scope> </dependency> <dependency> diff --git a/kafka-ui-serde-api/pom.xml b/kafka-ui-serde-api/pom.xml index 3b2b9f55fa9..b9aca207e58 100644 --- a/kafka-ui-serde-api/pom.xml +++ b/kafka-ui-serde-api/pom.xml @@ -13,10 +13,12 @@ <build> <plugins> <plugin> + <groupId>org.apache.maven.plugins</groupId> <artifactId>maven-install-plugin</artifactId> <version>2.5.2</version> </plugin> <plugin> + <groupId>org.apache.maven.plugins</groupId> <artifactId>maven-jar-plugin</artifactId> <version>3.0.2</version> </plugin> diff --git a/pom.xml b/pom.xml index ecb707e1da0..32bb407fff0 100644 --- a/pom.xml +++ b/pom.xml @@ -14,41 +14,54 @@ <maven.compiler.release>17</maven.compiler.release> <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding> - <spring-boot.version>2.7.5</spring-boot.version> - <jackson-databind-nullable.version>0.2.2</jackson-databind-nullable.version> - <org.mapstruct.version>1.4.2.Final</org.mapstruct.version> - <org.projectlombok.version>1.18.20</org.projectlombok.version> - <org.projectlombok.e2e-checks.version>1.18.20</org.projectlombok.e2e-checks.version> + <frontend-generated-sources-directory>..//kafka-ui-react-app/src/generated-sources + </frontend-generated-sources-directory> + <sonar.organization>provectus</sonar.organization> + <sonar.host.url>https://sonarcloud.io</sonar.host.url> <git.revision>latest</git.revision> - <kafka-clients.version>3.2.0</kafka-clients.version> + + <!-- Dependency versions --> + <antlr4-maven-plugin.version>4.7.1</antlr4-maven-plugin.version> + <apache.commons.version>2.11.1</apache.commons.version> + <assertj.version>3.19.0</assertj.version> + <avro.version>1.11.1</avro.version> + <byte-buddy.version>1.12.18</byte-buddy.version> + <confluent.version>7.3.0</confluent.version> + <datasketches-java.version>3.1.0</datasketches-java.version> + <groovy.version>3.0.13</groovy.version> + <jackson-databind-nullable.version>0.2.4</jackson-databind-nullable.version> + <jackson.version>2.14.0</jackson.version> + <kafka-clients.version>3.3.1</kafka-clients.version> + <netty.version>4.1.85.Final</netty.version> + <org.mapstruct.version>1.4.2.Final</org.mapstruct.version> + <org.projectlombok.version>1.18.24</org.projectlombok.version> + <protobuf-java.version>3.21.9</protobuf-java.version> + <reactor-netty.version>1.1.0</reactor-netty.version> + <scala-lang.library.version>2.13.9</scala-lang.library.version> + <snakeyaml.version>1.33</snakeyaml.version> + <spring-boot.version>2.7.5</spring-boot.version> + <spring-security.version>5.7.5</spring-security.version> + + <!-- Test dependency versions --> + <junit.version>5.9.1</junit.version> + <mockito.version>4.8.1</mockito.version> + <okhttp3.mockwebserver.version>4.10.0</okhttp3.mockwebserver.version> + <testcontainers.version>1.17.5</testcontainers.version> + + <!-- Frontend dependency versions --> <node.version>v16.15.0</node.version> <pnpm.version>v7.4.0</pnpm.version> + + <!-- Plugin versions --> <fabric8-maven-plugin.version>0.40.2</fabric8-maven-plugin.version> <frontend-maven-plugin.version>1.12.1</frontend-maven-plugin.version> - <maven-compiler-plugin.version>3.10.1</maven-compiler-plugin.version> <maven-clean-plugin.version>3.2.0</maven-clean-plugin.version> + <maven-compiler-plugin.version>3.10.1</maven-compiler-plugin.version> <maven-resources-plugin.version>3.2.0</maven-resources-plugin.version> <maven-surefire-plugin.version>2.22.2</maven-surefire-plugin.version> <openapi-generator-maven-plugin.version>4.3.0</openapi-generator-maven-plugin.version> - <swagger-annotations.version>1.6.0</swagger-annotations.version> <springdoc-openapi-webflux-ui.version>1.2.32</springdoc-openapi-webflux-ui.version> - <avro.version>1.11.0</avro.version> - <confluent.version>7.3.0</confluent.version> - <apache.commons.version>2.11.1</apache.commons.version> - <test.containers.version>1.17.1</test.containers.version> - <okhttp3.mockwebserver.version>4.10.0</okhttp3.mockwebserver.version> - <junit-jupiter-engine.version>5.7.2</junit-jupiter-engine.version> - <mockito.version>4.8.1</mockito.version> - <byte-buddy.version>1.12.18</byte-buddy.version> - <assertj.version>3.19.0</assertj.version> - <antlr4-maven-plugin.version>4.7.1</antlr4-maven-plugin.version> - <groovy.version>3.0.13</groovy.version> - <datasketches-java.version>3.1.0</datasketches-java.version> - - <frontend-generated-sources-directory>..//kafka-ui-react-app/src/generated-sources - </frontend-generated-sources-directory> - <sonar.organization>provectus</sonar.organization> - <sonar.host.url>https://sonarcloud.io</sonar.host.url> + <swagger-annotations.version>1.6.0</swagger-annotations.version> </properties> <repositories> @@ -86,6 +99,73 @@ </pluginRepository> </pluginRepositories> + <dependencyManagement> + <dependencies> + <dependency> + <groupId>org.springframework.boot</groupId> + <artifactId>spring-boot-dependencies</artifactId> + <version>${spring-boot.version}</version> + <type>pom</type> + <scope>import</scope> + </dependency> + <dependency> + <groupId>org.springframework.security</groupId> + <artifactId>spring-security-bom</artifactId> + <version>${spring-security.version}</version> + <type>pom</type> + <scope>import</scope> + </dependency> + <dependency> + <groupId>io.netty</groupId> + <artifactId>netty-bom</artifactId> + <version>${netty.version}</version> + <type>pom</type> + <scope>import</scope> + </dependency> + <dependency> + <groupId>com.fasterxml.jackson</groupId> + <artifactId>jackson-bom</artifactId> + <version>${jackson.version}</version> + <type>pom</type> + <scope>import</scope> + </dependency> + <dependency> + <groupId>org.scala-lang</groupId> + <artifactId>scala-library</artifactId> + <version>${scala-lang.library.version}</version> + </dependency> + <dependency> + <groupId>org.yaml</groupId> + <artifactId>snakeyaml</artifactId> + <version>${snakeyaml.version}</version> + </dependency> + <dependency> + <groupId>com.google.protobuf</groupId> + <artifactId>protobuf-java</artifactId> + <version>${protobuf-java.version}</version> + </dependency> + <dependency> + <groupId>io.projectreactor.netty</groupId> + <artifactId>reactor-netty-http</artifactId> + <version>${reactor-netty.version}</version> + </dependency> + <dependency> + <groupId>org.junit</groupId> + <artifactId>junit-bom</artifactId> + <version>${junit.version}</version> + <type>pom</type> + <scope>import</scope> + </dependency> + <dependency> + <groupId>org.testcontainers</groupId> + <artifactId>testcontainers-bom</artifactId> + <version>${testcontainers.version}</version> + <type>pom</type> + <scope>import</scope> + </dependency> + </dependencies> + </dependencyManagement> + <build> <pluginManagement> <plugins>
null
test
val
2022-11-14T17:02:34
"2022-10-07T08:47:40Z"
Haarolean
train
provectus/kafka-ui/2909_2931
provectus/kafka-ui
provectus/kafka-ui/2909
provectus/kafka-ui/2931
[ "connected" ]
f47a0382f73fc88b778d87492d09bc9b8427da1c
818149977cc234bf18ba9eeabb00926c16610bc4
[]
[ "why do we need this annotation?", "why do we need this annotation?", "extra space", "getVisibleSummaryCellL -> getVisibleSummaryCells", "can we simplify this way?\nreturn getEnabledColumnHeaders();", "really enabled?", "here and at 39 line we need to specify element.getSearchCriteria() + \" isVisibleOrEnabled()\" as we used in com.provectus.kafka.ui.suite.topics.TopicActionsTests.checkTopicListElements, because if some element(s) will fail assertion we'll know which one(s) only this way", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "Fixed." ]
"2022-11-14T13:38:22Z"
[ "scope/QA", "scope/AQA" ]
[e2e] Checking the existing Broker's profile in a cluster
Autotest implementation for: https://app.qase.io/case/KAFKAUI-85 Description: Checking the information of used Broker in a cluster Pre-conditions: Login to Kafka-ui Open the cluster Post-conditions: Empty value Steps: 1 Press the Brokers component 2 Click on Broker Expected result: 1 All the existing brokers should be displayed with different rows 2 The Broker's profile should open with following information: Segment Size, Segment Count, Port, Host, Name, Error, Topics, Partitions.
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersList.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java index 827a663cc51..de733b668f9 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java @@ -16,7 +16,9 @@ public abstract class BasePage extends WebUtils { protected SelenideElement dotMenuBtn = $x("//button[@aria-label='Dropdown Toggle']"); protected SelenideElement alertHeader = $x("//div[@role='alert']//div[@role='heading']"); protected SelenideElement alertMessage = $x("//div[@role='alert']//div[@role='contentinfo']"); + protected String summaryCellLocator = "//div[contains(text(),'%s')]"; protected String tableElementNameLocator = "//tbody//a[contains(text(),'%s')]"; + protected String сolumnHeaderLocator = "//table//tr/th/div[text()='%s']"; protected void waitUntilSpinnerDisappear() { log.debug("\nwaitUntilSpinnerDisappear"); diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersDetails.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersDetails.java new file mode 100644 index 00000000000..8ae28c0944e --- /dev/null +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersDetails.java @@ -0,0 +1,56 @@ +package com.provectus.kafka.ui.pages.brokers; + +import static com.codeborne.selenide.Selenide.$x; + +import com.codeborne.selenide.Condition; +import com.codeborne.selenide.SelenideElement; +import com.provectus.kafka.ui.pages.BasePage; +import io.qameta.allure.Step; +import java.util.ArrayList; +import java.util.Arrays; +import java.util.List; +import java.util.stream.Collectors; +import java.util.stream.Stream; + +public class BrokersDetails extends BasePage { + + protected SelenideElement logDirectoriesTab = $x("//a[text()='Log directories']"); + protected SelenideElement metricsTab = $x("//a[text()='Metrics']"); + + @Step + public BrokersDetails waitUntilScreenReady() { + waitUntilSpinnerDisappear(); + Arrays.asList(logDirectoriesTab, metricsTab).forEach(element -> element.shouldBe(Condition.visible)); + return this; + } + + private List<SelenideElement> getVisibleColumnHeaders() { + return Stream.of("Name", "Topics", "Error", "Partitions") + .map(name -> $x(String.format(сolumnHeaderLocator, name))) + .collect(Collectors.toList()); + } + + private List<SelenideElement> getEnabledColumnHeaders() { + return Stream.of("Name", "Error") + .map(name -> $x(String.format(сolumnHeaderLocator, name))) + .collect(Collectors.toList()); + } + + private List<SelenideElement> getVisibleSummaryCells() { + return Stream.of("Segment Size", "Segment Count", "Port", "Host") + .map(name -> $x(String.format(summaryCellLocator, name))) + .collect(Collectors.toList()); + } + + @Step + public List<SelenideElement> getAllEnabledElements() { + return getEnabledColumnHeaders(); + } + + @Step + public List<SelenideElement> getAllVisibleElements() { + List<SelenideElement> visibleElements = new ArrayList<>(getVisibleSummaryCells()); + visibleElements.addAll(getVisibleColumnHeaders()); + return visibleElements; + } +} diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersList.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersList.java new file mode 100644 index 00000000000..0d9ee67492c --- /dev/null +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersList.java @@ -0,0 +1,32 @@ +package com.provectus.kafka.ui.pages.brokers; + +import static com.codeborne.selenide.Selenide.$x; + +import com.codeborne.selenide.Condition; +import com.codeborne.selenide.SelenideElement; +import com.provectus.kafka.ui.pages.BasePage; +import io.qameta.allure.Step; + +public class BrokersList extends BasePage { + + protected SelenideElement brokersListHeader = $x("//h1[text()='Brokers']"); + + @Step + public BrokersList waitUntilScreenReady() { + waitUntilSpinnerDisappear(); + brokersListHeader.shouldBe(Condition.visible); + return this; + } + + @Step + public boolean isBrokerVisible(String brokerId) { + tableGrid.shouldBe(Condition.visible); + return isVisible(getTableElement(brokerId)); + } + + @Step + public BrokersList openBroker(String brokerName) { + getTableElement(brokerName).shouldBe(Condition.enabled).click(); + return this; + } +} diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java index cd61bd72721..4ce2721d6fd 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java @@ -21,7 +21,6 @@ public class TopicsList extends BasePage { protected SelenideElement addTopicBtn = $x("//button[normalize-space(text()) ='Add a Topic']"); protected SelenideElement searchField = $x("//input[@placeholder='Search by Topic Name']"); protected SelenideElement showInternalRadioBtn = $x("//input[@name='ShowInternalTopics']"); - protected String сolumnHeaderLocator = "//table//tr/th/div[text()='%s']"; protected String actionButtonLocator = "//button[text()='%s']"; @Step diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java index e35cad9769a..35a0a06f7a5 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java @@ -1,5 +1,7 @@ package com.provectus.kafka.ui.utilities; +import static com.codeborne.selenide.Selenide.executeJavaScript; + import com.codeborne.selenide.Condition; import com.codeborne.selenide.SelenideElement; import com.codeborne.selenide.WebDriverRunner; @@ -7,44 +9,42 @@ import org.openqa.selenium.Keys; import org.openqa.selenium.interactions.Actions; -import static com.codeborne.selenide.Selenide.executeJavaScript; - @Slf4j public class WebUtils { - public static void clickByActions(SelenideElement element) { - log.debug("\nclickByActions: {}", element.getSearchCriteria()); - element.shouldBe(Condition.enabled); - new Actions(WebDriverRunner.getWebDriver()) - .moveToElement(element) - .click(element) - .perform(); - } + public static void clickByActions(SelenideElement element) { + log.debug("\nclickByActions: {}", element.getSearchCriteria()); + element.shouldBe(Condition.enabled); + new Actions(WebDriverRunner.getWebDriver()) + .moveToElement(element) + .click(element) + .perform(); + } - public static void clickByJavaScript(SelenideElement element) { - log.debug("\nclickByJavaScript: {}", element.getSearchCriteria()); - element.shouldBe(Condition.enabled); - String script = "arguments[0].click();"; - executeJavaScript(script, element); - } + public static void clickByJavaScript(SelenideElement element) { + log.debug("\nclickByJavaScript: {}", element.getSearchCriteria()); + element.shouldBe(Condition.enabled); + String script = "arguments[0].click();"; + executeJavaScript(script, element); + } - public static void clearByKeyboard(SelenideElement field) { - log.debug("\nclearByKeyboard: {}", field.getSearchCriteria()); - field.shouldBe(Condition.enabled).sendKeys(Keys.END); - field.sendKeys(Keys.chord(Keys.CONTROL + "a"), Keys.DELETE); - } + public static void clearByKeyboard(SelenideElement field) { + log.debug("\nclearByKeyboard: {}", field.getSearchCriteria()); + field.shouldBe(Condition.enabled).sendKeys(Keys.END); + field.sendKeys(Keys.chord(Keys.CONTROL + "a"), Keys.DELETE); + } - public static boolean isVisible(SelenideElement element) { - log.debug("\nisVisible: {}", element.getSearchCriteria()); - boolean isVisible = false; - try { - element.shouldBe(Condition.visible); - isVisible = true; - } catch (Throwable e) { - log.debug("{} is not visible", element.getSearchCriteria()); - } - return isVisible; + public static boolean isVisible(SelenideElement element) { + log.debug("\nisVisible: {}", element.getSearchCriteria()); + boolean isVisible = false; + try { + element.shouldBe(Condition.visible); + isVisible = true; + } catch (Throwable e) { + log.debug("{} is not visible", element.getSearchCriteria()); } + return isVisible; + } public static boolean isEnabled(SelenideElement element) { log.debug("\nisEnabled: {}", element.getSearchCriteria());
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java index 5b9533c30ec..279704b8b95 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java @@ -3,6 +3,8 @@ import com.provectus.kafka.ui.helpers.ApiHelper; import com.provectus.kafka.ui.pages.NaviSideBar; import com.provectus.kafka.ui.pages.TopPanel; +import com.provectus.kafka.ui.pages.brokers.BrokersDetails; +import com.provectus.kafka.ui.pages.brokers.BrokersList; import com.provectus.kafka.ui.pages.connector.ConnectorCreateForm; import com.provectus.kafka.ui.pages.connector.ConnectorDetails; import com.provectus.kafka.ui.pages.connector.KafkaConnectList; @@ -32,4 +34,6 @@ public abstract class Facade { protected ConsumersList consumersList = new ConsumersList(); protected NaviSideBar naviSideBar = new NaviSideBar(); protected TopPanel topPanel = new TopPanel(); + protected BrokersList brokersList = new BrokersList(); + protected BrokersDetails brokersDetails = new BrokersDetails(); } diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java new file mode 100644 index 00000000000..5a493438f98 --- /dev/null +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java @@ -0,0 +1,42 @@ +package com.provectus.kafka.ui.suite.brokers; + +import static com.provectus.kafka.ui.pages.NaviSideBar.SideMenuOption.BROKERS; +import static org.assertj.core.api.Assertions.assertThat; + +import com.codeborne.selenide.Condition; +import com.provectus.kafka.ui.base.BaseTest; +import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.AutomationStatus; +import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.Suite; +import com.provectus.kafka.ui.utilities.qaseIoUtils.enums.Status; +import io.qase.api.annotation.CaseId; +import org.assertj.core.api.SoftAssertions; +import org.junit.jupiter.api.DisplayName; +import org.junit.jupiter.api.Test; + +public class BrokersTests extends BaseTest { + private static final String SUITE_TITLE = "Brokers"; + private static final long SUITE_ID = 1; + + @DisplayName("Checking the existing Broker's profile in a cluster") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(85) + @Test + public void checkExistingBrokersInCluster(){ + naviSideBar + .openSideMenu(BROKERS); + brokersList + .waitUntilScreenReady(); + assertThat(brokersList.isBrokerVisible("1")).as("isBrokerVisible()").isTrue(); + brokersList + .openBroker("1"); + brokersDetails + .waitUntilScreenReady(); + SoftAssertions softly = new SoftAssertions(); + brokersDetails.getAllVisibleElements().forEach(element -> softly.assertThat(element.is(Condition.visible)) + .as(element.getSearchCriteria() + " isVisible()").isTrue()); + brokersDetails.getAllEnabledElements().forEach(element -> softly.assertThat(element.is(Condition.enabled)) + .as(element.getSearchCriteria() + " isEnabled()").isTrue()); + softly.assertAll(); + } +}
train
val
2022-11-14T11:53:56
"2022-11-08T14:06:55Z"
ArthurNiedial
train
provectus/kafka-ui/2919_2939
provectus/kafka-ui
provectus/kafka-ui/2919
provectus/kafka-ui/2939
[ "connected" ]
93af20c46cacef03fbd06eceebe3a6caffda7b0d
d1013b69dadf9b5ff5037f3ab450b22bb11213cf
[]
[ "you're getting text from le element found by that text? no reason in this method, we need or to find filter name by another locator and assert text or assert that element found by text is visible", "1 we agreed use assertions only inside test classes\n2I meant create this method with modal element(s) (for example use addFilterCodeModalTitle) shouldBe(visible) only\n3 do not forget that we're waiting enabled and disabled elements are visible too\n4 looks according to AC disabled element is only addFilterBtnAddFilterMdl, so we can create isAddFilterBtnAddFilterModalEnable() method instead of getAllAddFilterModalDisabledElements()\n", "Done", "I refactored assertion block of code, and hope that I understood you correctly" ]
"2022-11-15T09:45:09Z"
[ "scope/AQA" ]
[e2e] Checking Filter adding within Topics/Topic/Messages with filled required fields
Autotest implementation for: https://app.qase.io/project/KAFKAUI?view=1&suite=5&previewMode=side&case=12 Description: Checking Filter adding within Topics/Topic profile/Messages tab Preconditions: Login to Kafka-ui application Open the 'Local' section Select the 'Topics' Steps: 1 Open any Topic's profile 2 Turn to 'Messages' tab 3 Press 'Add Filters' Expected result 'Add filter ' pop up should open with following information: ''Saved filters' link, 'Filter code' field, 'Save this filter' checkbox (unchecked by default), 'Display name' input field, 'Cancel' active and 'Add' disabled by default buttons 4 Fill 'Filter code' input field Expected result 'Add' button should become active 5 Press 'Add' button Expected result: 1 'Add filter' pop up should disappear 2 New filter should appear near '+Add Filters' button 3 Corresponding data should display
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java index 9ac485c26b7..e968d4f9372 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java @@ -13,6 +13,7 @@ import com.provectus.kafka.ui.utilities.WaitUtils; import io.qameta.allure.Step; import java.util.ArrayList; +import java.util.Arrays; import java.util.List; import lombok.experimental.ExtensionMethod; import org.openqa.selenium.By; @@ -24,6 +25,14 @@ public class TopicDetails extends BasePage { protected SelenideElement messageAmountCell = $x("//tbody/tr/td[5]"); protected SelenideElement overviewTab = $x("//a[contains(text(),'Overview')]"); protected SelenideElement messagesTab = $x("//a[contains(text(),'Messages')]"); + protected SelenideElement addFiltersBtn = $x("//button[text()='Add Filters']"); + protected SelenideElement savedFiltersField = $x("//div[text()='Saved Filters']"); + protected SelenideElement addFilterCodeModalTitle = $x("//label[text()='Filter code']"); + protected SelenideElement addFilterCodeInput = $x("//div[@id='ace-editor']//textarea"); + protected SelenideElement saveThisFilterCheckBoxAddFilterMdl = $x("//input[@name='saveFilter']"); + protected SelenideElement displayNameInputAddFilterMdl = $x("//input[@placeholder='Enter Name']"); + protected SelenideElement cancelBtnAddFilterMdl = $x("//button[text()='Cancel']"); + protected SelenideElement addFilterBtnAddFilterMdl = $x("//button[text()='Add filter']"); protected SelenideElement editSettingsMenu = $x("//li[@role][contains(text(),'Edit settings')]"); protected SelenideElement removeTopicBtn = $x("//ul[@role='menu']//div[contains(text(),'Remove Topic')]"); protected SelenideElement confirmBtn = $x("//div[@role='dialog']//button[contains(text(),'Confirm')]"); @@ -34,6 +43,7 @@ public class TopicDetails extends BasePage { protected ElementsCollection messageGridItems = $$x("//tbody//tr"); protected String consumerIdLocator = "//a[@title='%s']"; protected String topicHeaderLocator = "//h1[contains(text(),'%s')]"; + protected String filterNameLocator = "//*[@data-testid='activeSmartFilter']"; @Step public TopicDetails waitUntilScreenReady() { @@ -106,6 +116,58 @@ public TopicDetails clickProduceMessageBtn() { return this; } + @Step + public TopicDetails clickMessagesAddFiltersBtn() { + addFiltersBtn.shouldBe(Condition.enabled).click(); + return this; + } + + @Step + public TopicDetails waitUntilAddFiltersMdlVisible() { + addFilterCodeModalTitle.shouldBe(Condition.visible); + return this; + } + + @Step + public TopicDetails clickAddFilterBtnAddFilterMdl() { + addFilterBtnAddFilterMdl.shouldBe(Condition.enabled).click(); + addFilterCodeModalTitle.shouldBe(Condition.hidden); + return this; + } + + @Step + public TopicDetails setFilterCodeFieldAddFilterMdl(String filterCode) { + addFilterCodeInput.shouldBe(Condition.enabled).sendKeys(filterCode); + return this; + } + + @Step + public boolean isSaveThisFilterCheckBoxSelected() { + return isSelected(saveThisFilterCheckBoxAddFilterMdl); + } + + @Step + public boolean isAddFilterBtnAddFilterMdlEnabled() { + return isEnabled(addFilterBtnAddFilterMdl); + } + + @Step + public String getFilterName() { + return $x(filterNameLocator).getText(); + } + + public List<SelenideElement> getAllAddFilterModalVisibleElements() { + return Arrays.asList(savedFiltersField, displayNameInputAddFilterMdl, addFilterBtnAddFilterMdl, cancelBtnAddFilterMdl); + } + + public List<SelenideElement> getAllAddFilterModalEnabledElements() { + return Arrays.asList(displayNameInputAddFilterMdl, cancelBtnAddFilterMdl); + } + + public List<SelenideElement> getAllAddFilterModalDisabledElements() { + return Arrays.asList(addFilterBtnAddFilterMdl); + } + @Step public TopicDetails openConsumerGroup(String consumerId) { $x(String.format(consumerIdLocator, consumerId)).click(); @@ -146,6 +208,23 @@ public TopicDetails.MessageGridItem getRandomMessage() { return getMessage(nextInt(initItems().size() - 1)); } + public enum TopicMenu { + OVERVIEW("Overview"), + MESSAGES("Messages"), + CONSUMERS("Consumers"), + SETTINGS("Settings"); + + private final String value; + + TopicMenu(String value) { + this.value = value; + } + + public String toString() { + return value; + } + } + public static class MessageGridItem extends BasePage { private final SelenideElement element; @@ -211,21 +290,4 @@ public MessageGridItem clickSaveAsFile() { return this; } } - - public enum TopicMenu { - OVERVIEW("Overview"), - MESSAGES("Messages"), - CONSUMERS("Consumers"), - SETTINGS("Settings"); - - private final String value; - - TopicMenu(String value) { - this.value = value; - } - - public String toString() { - return value; - } - } } diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java index 4ce2721d6fd..dcda2ce4dff 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java @@ -17,7 +17,7 @@ @ExtensionMethod(WaitUtils.class) public class TopicsList extends BasePage { - protected SelenideElement topicListHeader = $x("//h1[text()='Topics']"); + protected SelenideElement topicListHeader = $x("//*[text()='Topics']"); protected SelenideElement addTopicBtn = $x("//button[normalize-space(text()) ='Add a Topic']"); protected SelenideElement searchField = $x("//input[@placeholder='Search by Topic Name']"); protected SelenideElement showInternalRadioBtn = $x("//input[@name='ShowInternalTopics']"); diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java index 35a0a06f7a5..3ba6d053ecd 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WebUtils.java @@ -57,4 +57,16 @@ public static boolean isEnabled(SelenideElement element) { } return isEnabled; } + + public static boolean isSelected(SelenideElement element) { + log.debug("\nisSelected: {}", element.getSearchCriteria()); + boolean isSelected = false; + try { + element.shouldBe(Condition.selected); + isSelected = true; + } catch (Throwable e) { + log.debug("{} is not selected", element.getSearchCriteria()); + } + return isSelected; + } }
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicActionsTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java similarity index 80% rename from kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicActionsTests.java rename to kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java index 3bc6dc07b4a..c342a216074 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicActionsTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java @@ -31,11 +31,11 @@ import org.junit.jupiter.api.TestInstance; @TestInstance(TestInstance.Lifecycle.PER_CLASS) -public class TopicActionsTests extends BaseTest { +public class TopicsTests extends BaseTest { private static final long SUITE_ID = 2; private static final String SUITE_TITLE = "Topics"; private static final Topic TOPIC_TO_CREATE = new Topic() - .setName("new-topic-"+ randomAlphabetic(5)) + .setName("new-topic-" + randomAlphabetic(5)) .setPartitions("1") .setCustomParameterType(COMPRESSION_TYPE) .setCustomParameterValue("producer") @@ -83,8 +83,10 @@ public void createTopic() { .waitUntilScreenReady() .openTopic(TOPIC_TO_CREATE.getName()); SoftAssertions softly = new SoftAssertions(); - softly.assertThat(topicDetails.isTopicHeaderVisible(TOPIC_TO_CREATE.getName())).as("isTopicHeaderVisible()").isTrue(); - softly.assertThat(topicDetails.getCleanUpPolicy()).as("getCleanUpPolicy()").isEqualTo(TOPIC_TO_CREATE.getCleanupPolicyValue().toString()); + softly.assertThat(topicDetails.isTopicHeaderVisible(TOPIC_TO_CREATE.getName())).as("isTopicHeaderVisible()") + .isTrue(); + softly.assertThat(topicDetails.getCleanUpPolicy()).as("getCleanUpPolicy()") + .isEqualTo(TOPIC_TO_CREATE.getCleanupPolicyValue().toString()); softly.assertThat(topicDetails.getPartitions()).as("getPartitions()").isEqualTo(TOPIC_TO_CREATE.getPartitions()); softly.assertAll(); naviSideBar @@ -132,10 +134,14 @@ public void updateTopic() { .openDotMenu() .clickEditSettingsMenu(); SoftAssertions softly = new SoftAssertions(); - softly.assertThat(topicCreateEditForm.getCleanupPolicy()).as("getCleanupPolicy()").isEqualTo(TOPIC_FOR_UPDATE.getCleanupPolicyValue().getVisibleText()); - softly.assertThat(topicCreateEditForm.getTimeToRetain()).as("getTimeToRetain()").isEqualTo(TOPIC_FOR_UPDATE.getTimeToRetainData()); - softly.assertThat(topicCreateEditForm.getMaxSizeOnDisk()).as("getMaxSizeOnDisk()").isEqualTo(TOPIC_FOR_UPDATE.getMaxSizeOnDisk().getVisibleText()); - softly.assertThat(topicCreateEditForm.getMaxMessageBytes()).as("getMaxMessageBytes()").isEqualTo(TOPIC_FOR_UPDATE.getMaxMessageBytes()); + softly.assertThat(topicCreateEditForm.getCleanupPolicy()).as("getCleanupPolicy()") + .isEqualTo(TOPIC_FOR_UPDATE.getCleanupPolicyValue().getVisibleText()); + softly.assertThat(topicCreateEditForm.getTimeToRetain()).as("getTimeToRetain()") + .isEqualTo(TOPIC_FOR_UPDATE.getTimeToRetainData()); + softly.assertThat(topicCreateEditForm.getMaxSizeOnDisk()).as("getMaxSizeOnDisk()") + .isEqualTo(TOPIC_FOR_UPDATE.getMaxSizeOnDisk().getVisibleText()); + softly.assertThat(topicCreateEditForm.getMaxMessageBytes()).as("getMaxMessageBytes()") + .isEqualTo(TOPIC_FOR_UPDATE.getMaxMessageBytes()); softly.assertAll(); } @@ -237,7 +243,7 @@ void checkCustomParametersWithinCreateNewTopic() { @AutomationStatus(status = Status.AUTOMATED) @CaseId(2) @Test - void checkTopicListElements(){ + void checkTopicListElements() { naviSideBar .openSideMenu(TOPICS); topicsList @@ -252,6 +258,45 @@ void checkTopicListElements(){ softly.assertAll(); } + @DisplayName("Filter adding within Topic") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(12) + @Test + void addingNewFilterWithinTopic() { + String topicName = "_schemas"; + String filterName = "123ABC"; + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .openTopic(topicName); + topicDetails + .openDetailsTab(TopicDetails.TopicMenu.MESSAGES) + .clickMessagesAddFiltersBtn() + .waitUntilAddFiltersMdlVisible(); + SoftAssertions softly = new SoftAssertions(); + topicDetails.getAllAddFilterModalVisibleElements().forEach(element -> + softly.assertThat(element.is(Condition.visible)) + .as(element.getSearchCriteria() + " isVisible()").isTrue()); + topicDetails.getAllAddFilterModalEnabledElements().forEach(element -> + softly.assertThat(element.is(Condition.enabled)) + .as(element.getSearchCriteria() + " isEnabled()").isTrue()); + topicDetails.getAllAddFilterModalDisabledElements().forEach(element -> + softly.assertThat(element.is(Condition.enabled)) + .as(element.getSearchCriteria() + " isEnabled()").isFalse()); + softly.assertThat(topicDetails.isSaveThisFilterCheckBoxSelected()).as("isSaveThisFilterCheckBoxSelected()") + .isFalse(); + softly.assertAll(); + topicDetails + .setFilterCodeFieldAddFilterMdl(filterName); + assertThat(topicDetails.isAddFilterBtnAddFilterMdlEnabled()).as("isMessagesAddFilterTabAddFilterBtnEnabled()") + .isTrue(); + topicDetails.clickAddFilterBtnAddFilterMdl(); + assertThat(topicDetails.getFilterName()).as("isFilterNameVisible(filterName)") + .isEqualTo(filterName); + } + @AfterAll public void afterAll() { TOPIC_LIST.forEach(topic -> apiHelper.deleteTopic(CLUSTER_NAME, topic.getName()));
test
val
2022-11-18T09:10:02
"2022-11-10T11:54:47Z"
Redbutton18
train
provectus/kafka-ui/2935_2940
provectus/kafka-ui
provectus/kafka-ui/2935
provectus/kafka-ui/2940
[ "connected" ]
0d0a24491586a499890897dac5f3546510f53cd5
5900f8e957f2abf078bb6db4552b186a9e21a0b5
[ "@nitin1677 thank you for issue!\r\nupdated docker image published (master tag) \r\ncan you please check on your setup?", "Updated the docker image with pull from master. No more error in docker logs. \r\n\r\n1. Topics without schema (Key serde: String, Value serde: String, Partitions: All items are selected) are able to pull the events and display in data grid.\r\n\r\n2. Topics with schema (Key serde: SchemaRegistry, Value serde: SchemaRegistry, Partitions: All items are selected) are not able to pull the events and result in data grid is, No messages found (2997 ms, 0 Bytes, 0 messages consumed)\r\n 2.1 From partitions list, select a partition (Partition#0) and data grid will display the events from select partition, 123 ms, 663 \r\n KB, 500 messages consumed.\r\n\r\nResult from docker logs when all items are selected for partitions:\r\n2022-11-15 13:33:09,908 INFO [boundedElastic-5] o.a.k.c.c.KafkaConsumer: [Consumer clientId=kafka-ui-92fe475e-27e5-4c74-8381-46b86892c0bb, groupId=null] Seeking to offset 654588 for partition TEST-TOPIC_V6-39\r\n2022-11-15 13:33:09,908 INFO [boundedElastic-5] o.a.k.c.c.KafkaConsumer: [Consumer clientId=kafka-ui-92fe475e-27e5-4c74-8381-46b86892c0bb, groupId=null] Seeking to offset 704717 for partition TEST-TOPIC_V6-8\r\n2022-11-15 13:33:10,908 INFO [boundedElastic-5] c.p.k.u.e.ForwardRecordEmitter: 0 records polled\r\n2022-11-15 13:33:11,914 INFO [boundedElastic-5] c.p.k.u.e.ForwardRecordEmitter: 0 records polled\r\n2022-11-15 13:33:12,916 INFO [boundedElastic-5] c.p.k.u.e.ForwardRecordEmitter: 0 records polled\r\n2022-11-15 13:33:12,917 INFO [boundedElastic-5] c.p.k.u.e.ForwardRecordEmitter: Polling finished\r\n2022-11-15 13:33:12,917 INFO [boundedElastic-5] o.a.k.c.m.Metrics: Metrics scheduler closed\r\n2022-11-15 13:33:12,917 INFO [boundedElastic-5] o.a.k.c.m.Metrics: Closing reporter org.apache.kafka.common.metrics.JmxReporter\r\n2022-11-15 13:33:12,917 INFO [boundedElastic-5] o.a.k.c.m.Metrics: Metrics reporters closed\r\n\r\n\r\n", "@nitin1677 thank you for checking\r\nlooks like snappy issue is solved, as for 2. that you mention - looks like it is separate issue, lets create ticket for that.\r\n\r\nTo finally verify that snappy works fine - can you please run topic analysis (Statistics tab) on both SR and non-SR topic?", "Yes, the snappy error is resolved and UI is functional to retrieve the events in data grid.\r\n\r\nRan topics analysis successfully for both Schema-Enabled topics and Topics without Schema.", "@nitin1677 thank you for confirmation!" ]
[]
"2022-11-15T10:34:14Z"
[ "type/bug", "scope/backend", "status/accepted", "status/confirmed" ]
Message tab infinite polling
<!-- Don't forget to check for existing issues/discussions regarding your proposal. We might already have it. https://github.com/provectus/kafka-ui/issues https://github.com/provectus/kafka-ui/discussions --> **Describe the bug** <!--(A clear and concise description of what the bug is.)--> After latest pull from master today, kafka-ui broke and no events are being displayed in ui. Switch to Kafka Cluster > topics > navigate to topic. Topic overview tab shows correct information. When navigate to messages tab, none of the events displayed in data grid and result is: Polling 0 ms 0 Bytes 0 messages consumed From docker logs, error is: 2022-11-15 00:06:24,483 ERROR [boundedElastic-4] r.c.s.Schedulers: Scheduler worker in group main failed with an uncaught exception java.lang.NoClassDefFoundError: Could not initialize class org.xerial.snappy.pure.UnsafeUtil The complete error stack is below. **Set up** <!-- How do you run the app? Please provide as much info as possible: 1. App version (docker image version or check commit hash in the top left corner in UI) 2. Helm chart version, if you use one 3. Any IAAC configs We might close the issue without further explanation if you don't provide such information. --> docker pull provectuslabs/kafka-ui:master 11.14.2022 11:03:06 ( [8181499](https://github.com/provectus/kafka-ui/commit/8181499) ) **Steps to Reproduce** <!-- We'd like you to provide an example setup (via docker-compose, helm, etc.) to reproduce the problem, especially with a complex setups. --> Steps to reproduce the behavior: 1. docker pull provectuslabs/kafka-ui:master **Expected behavior** <!-- (A clear and concise description of what you expected to happen) --> Events from kafka topics to be displayed in data grid. **Screenshots** <!-- (If applicable, add screenshots to help explain your problem) --> **Additional context** <!-- (Add any other context about the problem here) --> 2022-11-15 00:06:24,224 INFO [boundedElastic-4] o.a.k.c.Metadata: [Consumer clientId=kafka-ui-25f7d8b6-b232-4643-b483-40b0df3f28e7, groupId=null] Cluster ID: Zp6dCNprQFWXYfHx6rwlkw 2022-11-15 00:06:24,274 INFO [boundedElastic-4] o.a.k.c.c.KafkaConsumer: [Consumer clientId=kafka-ui-25f7d8b6-b232-4643-b483-40b0df3f28e7, groupId=null] Subscribed to partition(s): TEST-TOPIC_V3-1, TEST-TOPIC_V3-2, TEST-TOPIC_V3-0 2022-11-15 00:06:24,275 INFO [boundedElastic-4] o.a.k.c.c.KafkaConsumer: [Consumer clientId=kafka-ui-25f7d8b6-b232-4643-b483-40b0df3f28e7, groupId=null] Seeking to offset 131339 for partition TEST-TOPIC_V3-1 2022-11-15 00:06:24,275 INFO [boundedElastic-4] o.a.k.c.c.KafkaConsumer: [Consumer clientId=kafka-ui-25f7d8b6-b232-4643-b483-40b0df3f28e7, groupId=null] Seeking to offset 134131 for partition TEST-TOPIC_V3-2 2022-11-15 00:06:24,275 INFO [boundedElastic-4] o.a.k.c.c.KafkaConsumer: [Consumer clientId=kafka-ui-25f7d8b6-b232-4643-b483-40b0df3f28e7, groupId=null] Seeking to offset 132305 for partition TEST-TOPIC_V3-0 2022-11-15 00:06:24,480 INFO [boundedElastic-4] o.a.k.c.m.Metrics: Metrics scheduler closed 2022-11-15 00:06:24,480 INFO [boundedElastic-4] o.a.k.c.m.Metrics: Closing reporter org.apache.kafka.common.metrics.JmxReporter 2022-11-15 00:06:24,480 INFO [boundedElastic-4] o.a.k.c.m.Metrics: Metrics reporters closed 2022-11-15 00:06:24,482 INFO [boundedElastic-4] o.a.k.c.u.AppInfoParser: App info kafka.consumer for kafka-ui-25f7d8b6-b232-4643-b483-40b0df3f28e7 unregistered 2022-11-15 00:06:24,483 WARN [boundedElastic-4] r.c.Exceptions: throwIfFatal detected a jvm fatal exception, which is thrown and logged below: java.lang.NoClassDefFoundError: Could not initialize class org.xerial.snappy.pure.UnsafeUtil at org.xerial.snappy.pure.SnappyRawDecompressor.getUnsignedByteSafe(SnappyRawDecompressor.java:323) at org.xerial.snappy.pure.SnappyRawDecompressor.readUncompressedLength(SnappyRawDecompressor.java:288) at org.xerial.snappy.pure.SnappyRawDecompressor.getUncompressedLength(SnappyRawDecompressor.java:42) at org.xerial.snappy.pure.PureJavaSnappy.uncompressedLength(PureJavaSnappy.java:238) at org.xerial.snappy.Snappy.uncompressedLength(Snappy.java:638) at org.xerial.snappy.SnappyInputStream.hasNextChunk(SnappyInputStream.java:435) at org.xerial.snappy.SnappyInputStream.read(SnappyInputStream.java:466) at java.base/java.io.DataInputStream.readByte(DataInputStream.java:271) at org.apache.kafka.common.utils.ByteUtils.readUnsignedVarint(ByteUtils.java:170) at org.apache.kafka.common.utils.ByteUtils.readVarint(ByteUtils.java:205) at org.apache.kafka.common.record.DefaultRecord.readFrom(DefaultRecord.java:281) at org.apache.kafka.common.record.DefaultRecordBatch$2.doReadRecord(DefaultRecordBatch.java:296) at org.apache.kafka.common.record.DefaultRecordBatch$StreamRecordIterator.readNext(DefaultRecordBatch.java:638) at org.apache.kafka.common.record.DefaultRecordBatch$RecordIterator.next(DefaultRecordBatch.java:603) at org.apache.kafka.common.record.DefaultRecordBatch$RecordIterator.next(DefaultRecordBatch.java:572) at org.apache.kafka.clients.consumer.internals.Fetcher$CompletedFetch.nextFetchedRecord(Fetcher.java:1618) at org.apache.kafka.clients.consumer.internals.Fetcher$CompletedFetch.fetchRecords(Fetcher.java:1653) at org.apache.kafka.clients.consumer.internals.Fetcher$CompletedFetch.access$1900(Fetcher.java:1494) at org.apache.kafka.clients.consumer.internals.Fetcher.fetchRecords(Fetcher.java:716) at org.apache.kafka.clients.consumer.internals.Fetcher.collectFetch(Fetcher.java:682) at org.apache.kafka.clients.consumer.KafkaConsumer.pollForFetches(KafkaConsumer.java:1313) at org.apache.kafka.clients.consumer.KafkaConsumer.poll(KafkaConsumer.java:1242) at org.apache.kafka.clients.consumer.KafkaConsumer.poll(KafkaConsumer.java:1215) at com.provectus.kafka.ui.emitter.AbstractEmitter.poll(AbstractEmitter.java:40) at com.provectus.kafka.ui.emitter.AbstractEmitter.poll(AbstractEmitter.java:34) at com.provectus.kafka.ui.emitter.ForwardRecordEmitter.accept(ForwardRecordEmitter.java:45) at com.provectus.kafka.ui.emitter.ForwardRecordEmitter.accept(ForwardRecordEmitter.java:14) at reactor.core.publisher.FluxCreate.subscribe(FluxCreate.java:95) at reactor.core.publisher.InternalFluxOperator.subscribe(InternalFluxOperator.java:62) at reactor.core.publisher.FluxSubscribeOn$SubscribeOnSubscriber.run(FluxSubscribeOn.java:194) at reactor.core.scheduler.WorkerTask.call(WorkerTask.java:84) at reactor.core.scheduler.WorkerTask.call(WorkerTask.java:37) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:833) 2022-11-15 00:06:24,483 WARN [boundedElastic-4] r.c.Exceptions: throwIfFatal detected a jvm fatal exception, which is thrown and logged below: java.lang.NoClassDefFoundError: Could not initialize class org.xerial.snappy.pure.UnsafeUtil at org.xerial.snappy.pure.SnappyRawDecompressor.getUnsignedByteSafe(SnappyRawDecompressor.java:323) at org.xerial.snappy.pure.SnappyRawDecompressor.readUncompressedLength(SnappyRawDecompressor.java:288) at org.xerial.snappy.pure.SnappyRawDecompressor.getUncompressedLength(SnappyRawDecompressor.java:42) at org.xerial.snappy.pure.PureJavaSnappy.uncompressedLength(PureJavaSnappy.java:238) at org.xerial.snappy.Snappy.uncompressedLength(Snappy.java:638) at org.xerial.snappy.SnappyInputStream.hasNextChunk(SnappyInputStream.java:435) at org.xerial.snappy.SnappyInputStream.read(SnappyInputStream.java:466) at java.base/java.io.DataInputStream.readByte(DataInputStream.java:271) at org.apache.kafka.common.utils.ByteUtils.readUnsignedVarint(ByteUtils.java:170) at org.apache.kafka.common.utils.ByteUtils.readVarint(ByteUtils.java:205) at org.apache.kafka.common.record.DefaultRecord.readFrom(DefaultRecord.java:281) at org.apache.kafka.common.record.DefaultRecordBatch$2.doReadRecord(DefaultRecordBatch.java:296) at org.apache.kafka.common.record.DefaultRecordBatch$StreamRecordIterator.readNext(DefaultRecordBatch.java:638) at org.apache.kafka.common.record.DefaultRecordBatch$RecordIterator.next(DefaultRecordBatch.java:603) at org.apache.kafka.common.record.DefaultRecordBatch$RecordIterator.next(DefaultRecordBatch.java:572) at org.apache.kafka.clients.consumer.internals.Fetcher$CompletedFetch.nextFetchedRecord(Fetcher.java:1618) at org.apache.kafka.clients.consumer.internals.Fetcher$CompletedFetch.fetchRecords(Fetcher.java:1653) at org.apache.kafka.clients.consumer.internals.Fetcher$CompletedFetch.access$1900(Fetcher.java:1494) at org.apache.kafka.clients.consumer.internals.Fetcher.fetchRecords(Fetcher.java:716) at org.apache.kafka.clients.consumer.internals.Fetcher.collectFetch(Fetcher.java:682) at org.apache.kafka.clients.consumer.KafkaConsumer.pollForFetches(KafkaConsumer.java:1313) at org.apache.kafka.clients.consumer.KafkaConsumer.poll(KafkaConsumer.java:1242) at org.apache.kafka.clients.consumer.KafkaConsumer.poll(KafkaConsumer.java:1215) at com.provectus.kafka.ui.emitter.AbstractEmitter.poll(AbstractEmitter.java:40) at com.provectus.kafka.ui.emitter.AbstractEmitter.poll(AbstractEmitter.java:34) at com.provectus.kafka.ui.emitter.ForwardRecordEmitter.accept(ForwardRecordEmitter.java:45) at com.provectus.kafka.ui.emitter.ForwardRecordEmitter.accept(ForwardRecordEmitter.java:14) at reactor.core.publisher.FluxCreate.subscribe(FluxCreate.java:95) at reactor.core.publisher.InternalFluxOperator.subscribe(InternalFluxOperator.java:62) at reactor.core.publisher.FluxSubscribeOn$SubscribeOnSubscriber.run(FluxSubscribeOn.java:194) at reactor.core.scheduler.WorkerTask.call(WorkerTask.java:84) at reactor.core.scheduler.WorkerTask.call(WorkerTask.java:37) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:833) 2022-11-15 00:06:24,483 ERROR [boundedElastic-4] r.c.s.Schedulers: Scheduler worker in group main failed with an uncaught exception java.lang.NoClassDefFoundError: Could not initialize class org.xerial.snappy.pure.UnsafeUtil at org.xerial.snappy.pure.SnappyRawDecompressor.getUnsignedByteSafe(SnappyRawDecompressor.java:323) at org.xerial.snappy.pure.SnappyRawDecompressor.readUncompressedLength(SnappyRawDecompressor.java:288) at org.xerial.snappy.pure.SnappyRawDecompressor.getUncompressedLength(SnappyRawDecompressor.java:42) at org.xerial.snappy.pure.PureJavaSnappy.uncompressedLength(PureJavaSnappy.java:238) at org.xerial.snappy.Snappy.uncompressedLength(Snappy.java:638) at org.xerial.snappy.SnappyInputStream.hasNextChunk(SnappyInputStream.java:435) at org.xerial.snappy.SnappyInputStream.read(SnappyInputStream.java:466) at java.base/java.io.DataInputStream.readByte(DataInputStream.java:271) at org.apache.kafka.common.utils.ByteUtils.readUnsignedVarint(ByteUtils.java:170) at org.apache.kafka.common.utils.ByteUtils.readVarint(ByteUtils.java:205) at org.apache.kafka.common.record.DefaultRecord.readFrom(DefaultRecord.java:281) at org.apache.kafka.common.record.DefaultRecordBatch$2.doReadRecord(DefaultRecordBatch.java:296) at org.apache.kafka.common.record.DefaultRecordBatch$StreamRecordIterator.readNext(DefaultRecordBatch.java:638) at org.apache.kafka.common.record.DefaultRecordBatch$RecordIterator.next(DefaultRecordBatch.java:603) at org.apache.kafka.common.record.DefaultRecordBatch$RecordIterator.next(DefaultRecordBatch.java:572) at org.apache.kafka.clients.consumer.internals.Fetcher$CompletedFetch.nextFetchedRecord(Fetcher.java:1618) at org.apache.kafka.clients.consumer.internals.Fetcher$CompletedFetch.fetchRecords(Fetcher.java:1653) at org.apache.kafka.clients.consumer.internals.Fetcher$CompletedFetch.access$1900(Fetcher.java:1494) at org.apache.kafka.clients.consumer.internals.Fetcher.fetchRecords(Fetcher.java:716) at org.apache.kafka.clients.consumer.internals.Fetcher.collectFetch(Fetcher.java:682) at org.apache.kafka.clients.consumer.KafkaConsumer.pollForFetches(KafkaConsumer.java:1313) at org.apache.kafka.clients.consumer.KafkaConsumer.poll(KafkaConsumer.java:1242) at org.apache.kafka.clients.consumer.KafkaConsumer.poll(KafkaConsumer.java:1215) at com.provectus.kafka.ui.emitter.AbstractEmitter.poll(AbstractEmitter.java:40) at com.provectus.kafka.ui.emitter.AbstractEmitter.poll(AbstractEmitter.java:34) at com.provectus.kafka.ui.emitter.ForwardRecordEmitter.accept(ForwardRecordEmitter.java:45) at com.provectus.kafka.ui.emitter.ForwardRecordEmitter.accept(ForwardRecordEmitter.java:14) at reactor.core.publisher.FluxCreate.subscribe(FluxCreate.java:95) at reactor.core.publisher.InternalFluxOperator.subscribe(InternalFluxOperator.java:62) at reactor.core.publisher.FluxSubscribeOn$SubscribeOnSubscriber.run(FluxSubscribeOn.java:194) at reactor.core.scheduler.WorkerTask.call(WorkerTask.java:84) at reactor.core.scheduler.WorkerTask.call(WorkerTask.java:37) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:833) 2022-11-15 00:06:25,123 DEBUG [parallel-3] c.p.k.u.s.ClustersStatisticsScheduler: Metrics updated for cluster: KAFKA_QA
[ "kafka-ui-api/Dockerfile" ]
[ "kafka-ui-api/Dockerfile" ]
[]
diff --git a/kafka-ui-api/Dockerfile b/kafka-ui-api/Dockerfile index c0c68ff0041..96e0eb414b1 100644 --- a/kafka-ui-api/Dockerfile +++ b/kafka-ui-api/Dockerfile @@ -1,5 +1,6 @@ FROM azul/zulu-openjdk-alpine:17 +RUN apk add --no-cache gcompat # need to make snappy codec work RUN addgroup -S kafkaui && adduser -S kafkaui -G kafkaui USER kafkaui
null
train
val
2022-11-15T11:22:16
"2022-11-15T02:27:46Z"
nitin1677
train
provectus/kafka-ui/2869_2941
provectus/kafka-ui
provectus/kafka-ui/2869
provectus/kafka-ui/2941
[ "connected" ]
9256b741bee0d5f1aec7f2e98019019b6e6c1cd2
7e82e7434ea2d555d881e03a230e5e21db0636d4
[ "The important precondition is to have \"show internal topics\" on", "Seems not fixed, unchecking is not available." ]
[ "I don't think this is the correct solution you just removed the props from the common component , this change will break the check when all the items are selected the header will not show the check.\r\n\r\nin `SelectRowHeader` debug onChange function\r\n\r\n```\r\nconst SelectRowHeader: React.FC<HeaderContext<unknown, unknown>> = ({\r\n table,\r\n}) => {\r\n const c = (a: unknown) => {\r\n console.log('something', table.getIsAllPageRowsSelected());\r\n table.getToggleAllPageRowsSelectedHandler()(a);\r\n };\r\n\r\n return (\r\n <IndeterminateCheckbox\r\n checked={table.getIsAllPageRowsSelected()}\r\n indeterminate={table.getIsSomePageRowsSelected()}\r\n onChange={c}\r\n />\r\n );\r\n};\r\n```\r\n\r\ncheck the result in the working case , vs non working case where the result in console stays false. we need to check Tanstack table documentation to see how to navigate around this problem", "I have thought we don't need default value on header checkbox", "`table.getIsAllRowsSelected()` yep this is the correct config", "We need to change this as well , to `getIsSomeRowsSelected`", "`getToggleAllRowsSelectedHandler()` since they all should rely on on Rows selected instead of a PageRows" ]
"2022-11-15T12:50:43Z"
[ "type/bug", "good first issue", "scope/frontend", "status/accepted", "status/confirmed" ]
Unchecking the already checked Topics from main checkbox is not available
**Describe the bug** Unchecking the already checked Topics from main checkbox is not working **Set up** https://www.kafka-ui.provectus.io/ **Steps to Reproduce** Steps to reproduce the behavior: 1. Navigate to Topics page 2. Check the main checkbox for all Topics 3. Make sure all topics are checked 4. Uncheck the main checkbox **Expected behavior** All the Topics should became unchecked **Screenshots** https://user-images.githubusercontent.com/104780608/199652821-94339eab-2552-47f7-933d-0f312b614211.mov **Additional context** <!-- (Add any other context about the problem here) -->
[ "kafka-ui-react-app/src/components/common/NewTable/SelectRowHeader.tsx" ]
[ "kafka-ui-react-app/src/components/common/NewTable/SelectRowHeader.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/common/NewTable/SelectRowHeader.tsx b/kafka-ui-react-app/src/components/common/NewTable/SelectRowHeader.tsx index 8121f3cd446..d33282f2838 100644 --- a/kafka-ui-react-app/src/components/common/NewTable/SelectRowHeader.tsx +++ b/kafka-ui-react-app/src/components/common/NewTable/SelectRowHeader.tsx @@ -6,9 +6,9 @@ const SelectRowHeader: React.FC<HeaderContext<unknown, unknown>> = ({ table, }) => ( <IndeterminateCheckbox - checked={table.getIsAllPageRowsSelected()} - indeterminate={table.getIsSomePageRowsSelected()} - onChange={table.getToggleAllPageRowsSelectedHandler()} + checked={table.getIsAllRowsSelected()} + indeterminate={table.getIsSomeRowsSelected()} + onChange={table.getToggleAllRowsSelectedHandler()} /> );
null
val
val
2022-11-22T13:40:34
"2022-11-03T05:29:57Z"
armenuikafka
train
provectus/kafka-ui/2932_2951
provectus/kafka-ui
provectus/kafka-ui/2932
provectus/kafka-ui/2951
[ "connected" ]
d1013b69dadf9b5ff5037f3ab450b22bb11213cf
9457d5ff2b7346d4dfd815a6bfcbb7612cab9acd
[]
[ "elementLocatorDdl -> ddlElementLocator", "here we can use primitive: Integer -> int", "getAllMarkedLines -> getMarkedLinesNumber", "selectVersionFromLeftDropDown - just left?\nversionNumberDd -> version", "remove pls", "extra space", "this method and it's object also aren't used\nwhy did we add them?", "u've just deleted deleteSchemaAvro??", "refactored to int \r\n", "done\r\n", "discussed ", "ye , sorry ", "done", "discussed ", "ye , sorry it my fail", "done ", "seems we can create method openSchemaVersionDdl()\nthen getVersionsNumberFromList() and after this assertion selectVersionFromDdl() - without schemaVersionDdl.shouldBe(Condition.enabled).click();", "done ", "can we not use duplicates this way?\nint latestVersion = schemaDetails\n.waitUntilScreenReady()\n.getLatestVersion();", "can we not use duplicates this way?\nint latestVersion = schemaCreateForm\n.waitUntilScreenReady()\n.openSchemaVersionDdl()\n.getVersionsNumberFromList();", "Compare versions is failed -> Versions number is not matched", "countOfElementsDd -> versionsNumberFromDdl", "done", "done", "done", "done\r\n" ]
"2022-11-18T11:01:20Z"
[ "scope/QA", "scope/AQA" ]
[e2e]Checking Compare Versions operation for Schema
**Autotest implementation for:** https://app.qase.io/case/KAFKAUI-44 **Description:** The purpose of this case is to check the Compare Versions operation for Schemas **Pre-conditions:** - Login to Kafka-ui - Navigate to Schema Registry - Open the Schema **Post-conditions:** Empty value **Steps:** 1. Press "Compare Versions" button **Expected result:** Compare page should open with 2 column 2. Press "Version *" dropdown **Expected result:** All the existing versions should be listed 3. Select different versions to compare **Expected result:** All the changes should be highlighted
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaCreateForm.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WaitUtils.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaCreateForm.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/schemas/SchemasTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java index de733b668f9..069b1b41dda 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java @@ -18,7 +18,7 @@ public abstract class BasePage extends WebUtils { protected SelenideElement alertMessage = $x("//div[@role='alert']//div[@role='contentinfo']"); protected String summaryCellLocator = "//div[contains(text(),'%s')]"; protected String tableElementNameLocator = "//tbody//a[contains(text(),'%s')]"; - protected String сolumnHeaderLocator = "//table//tr/th/div[text()='%s']"; + protected String columnHeaderLocator = "//table//tr/th/div[text()='%s']"; protected void waitUntilSpinnerDisappear() { log.debug("\nwaitUntilSpinnerDisappear"); diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersDetails.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersDetails.java index 8ae28c0944e..77fc9e0b41c 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersDetails.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersDetails.java @@ -26,13 +26,13 @@ public BrokersDetails waitUntilScreenReady() { private List<SelenideElement> getVisibleColumnHeaders() { return Stream.of("Name", "Topics", "Error", "Partitions") - .map(name -> $x(String.format(сolumnHeaderLocator, name))) + .map(name -> $x(String.format(columnHeaderLocator, name))) .collect(Collectors.toList()); } private List<SelenideElement> getEnabledColumnHeaders() { return Stream.of("Name", "Error") - .map(name -> $x(String.format(сolumnHeaderLocator, name))) + .map(name -> $x(String.format(columnHeaderLocator, name))) .collect(Collectors.toList()); } diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java index 316178d1d22..b86734b2dc6 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/connector/KafkaConnectList.java @@ -5,12 +5,9 @@ import com.codeborne.selenide.Condition; import com.codeborne.selenide.SelenideElement; import com.provectus.kafka.ui.pages.BasePage; -import com.provectus.kafka.ui.utilities.WaitUtils; import io.qameta.allure.Step; -import lombok.experimental.ExtensionMethod; -@ExtensionMethod(WaitUtils.class) public class KafkaConnectList extends BasePage { protected SelenideElement createConnectorBtn = $x("//button[contains(text(),'Create Connector')]"); diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaCreateForm.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaCreateForm.java index ec6455fcfa2..7d0d5c58a93 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaCreateForm.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaCreateForm.java @@ -1,6 +1,7 @@ package com.provectus.kafka.ui.pages.schema; import static com.codeborne.selenide.Selenide.$; +import static com.codeborne.selenide.Selenide.$$x; import static com.codeborne.selenide.Selenide.$x; import com.codeborne.selenide.Condition; @@ -9,6 +10,7 @@ import com.provectus.kafka.ui.api.model.SchemaType; import com.provectus.kafka.ui.pages.BasePage; import io.qameta.allure.Step; +import java.util.List; public class SchemaCreateForm extends BasePage { @@ -19,7 +21,10 @@ public class SchemaCreateForm extends BasePage { protected SelenideElement schemaTypeDdl = $x("//ul[@name='schemaType']"); protected SelenideElement compatibilityLevelList = $x("//ul[@name='compatibilityLevel']"); protected SelenideElement newSchemaTextArea = $x("//div[@id='newSchema']"); - protected String elementLocatorDdl = "//li[@value='%s']"; + protected SelenideElement schemaVersionDdl = $$x("//ul[@role='listbox']/li[text()='Version 2']").first(); + protected List<SelenideElement> visibleMarkers = $$x("//div[@class='ace_scroller']//div[contains(@class,'codeMarker')]"); + protected List<SelenideElement> elementsCompareVersionDdl = $$x("//ul[@role='listbox']/ul/li"); + protected String ddlElementLocator = "//li[@value='%s']"; @Step public SchemaCreateForm waitUntilScreenReady(){ @@ -43,7 +48,7 @@ public SchemaCreateForm setSchemaField(String text) { @Step public SchemaCreateForm selectSchemaTypeFromDropdown(SchemaType schemaType) { schemaTypeDdl.shouldBe(Condition.enabled).click(); - $x(String.format(elementLocatorDdl, schemaType.getValue())).shouldBe(Condition.visible).click(); + $x(String.format(ddlElementLocator, schemaType.getValue())).shouldBe(Condition.visible).click(); return this; } @@ -56,10 +61,32 @@ public SchemaCreateForm clickSubmitButton() { @Step public SchemaCreateForm selectCompatibilityLevelFromDropdown(CompatibilityLevel.CompatibilityEnum level) { compatibilityLevelList.shouldBe(Condition.enabled).click(); - $x(String.format(elementLocatorDdl, level.getValue())).shouldBe(Condition.visible).click(); + $x(String.format(ddlElementLocator, level.getValue())).shouldBe(Condition.visible).click(); return this; } + @Step + public SchemaCreateForm openSchemaVersionDdl(){ + schemaVersionDdl.shouldBe(Condition.enabled).click(); + return this; + } + + @Step + public int getVersionsNumberFromList(){ + return elementsCompareVersionDdl.size(); + } + + @Step + public SchemaCreateForm selectVersionFromDropDown(int versionNumberDd){ + $x(String.format(ddlElementLocator,versionNumberDd)).shouldBe(Condition.visible).click(); + return this; + } + + @Step + public int getMarkedLinesNumber(){ + return visibleMarkers.size(); + } + @Step public SchemaCreateForm setNewSchemaValue(String configJson) { newSchemaTextArea.shouldBe(Condition.visible).click(); diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaDetails.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaDetails.java index b9808d05405..ce211337099 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaDetails.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaDetails.java @@ -14,7 +14,9 @@ public class SchemaDetails extends BasePage { protected SelenideElement editSchemaBtn = $x("//button[contains(text(),'Edit Schema')]"); protected SelenideElement removeBtn = $x("//*[contains(text(),'Remove')]"); protected SelenideElement confirmBtn = $x("//div[@role='dialog']//button[contains(text(),'Confirm')]"); - protected SelenideElement schemaTypeDdl = $x("//h4[contains(text(),'Type')]/../p"); + protected SelenideElement schemaTypeField = $x("//h4[contains(text(),'Type')]/../p"); + protected SelenideElement latestVersionField = $x("//h4[contains(text(),'Latest version')]/../p"); + protected SelenideElement compareVersionBtn = $x("//button[text()='Compare Versions']"); protected String schemaHeaderLocator = "//h1[contains(text(),'%s')]"; @Step @@ -34,9 +36,14 @@ public boolean isSchemaHeaderVisible(String schemaName) { return isVisible($x(String.format(schemaHeaderLocator,schemaName))); } + @Step + public int getLatestVersion(){ + return Integer.parseInt(latestVersionField.getText()); + } + @Step public String getSchemaType() { - return schemaTypeDdl.getText(); + return schemaTypeField.getText(); } @Step @@ -45,6 +52,12 @@ public SchemaDetails openEditSchema(){ return this; } + @Step + public SchemaDetails openCompareVersionMenu(){ + compareVersionBtn.shouldBe(Condition.enabled).click(); + return this; + } + @Step public SchemaDetails removeSchema() { clickByJavaScript(dotMenuBtn); diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java index e968d4f9372..95c3e6fb23f 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java @@ -10,15 +10,12 @@ import com.codeborne.selenide.ElementsCollection; import com.codeborne.selenide.SelenideElement; import com.provectus.kafka.ui.pages.BasePage; -import com.provectus.kafka.ui.utilities.WaitUtils; import io.qameta.allure.Step; import java.util.ArrayList; import java.util.Arrays; import java.util.List; -import lombok.experimental.ExtensionMethod; import org.openqa.selenium.By; -@ExtensionMethod({WaitUtils.class}) public class TopicDetails extends BasePage { protected SelenideElement clearMessagesBtn = $x(("//div[contains(text(), 'Clear messages')]")); diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java index dcda2ce4dff..c32621bf1a5 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java @@ -5,16 +5,13 @@ import com.codeborne.selenide.Condition; import com.codeborne.selenide.SelenideElement; import com.provectus.kafka.ui.pages.BasePage; -import com.provectus.kafka.ui.utilities.WaitUtils; import io.qameta.allure.Step; import java.util.ArrayList; import java.util.Arrays; import java.util.List; import java.util.stream.Collectors; import java.util.stream.Stream; -import lombok.experimental.ExtensionMethod; -@ExtensionMethod(WaitUtils.class) public class TopicsList extends BasePage { protected SelenideElement topicListHeader = $x("//*[text()='Topics']"); @@ -56,13 +53,13 @@ private List<SelenideElement> getActionButtons() { private List<SelenideElement> getVisibleColumnHeaders() { return Stream.of("Replication Factor","Number of messages","Topic Name", "Partitions", "Out of sync replicas", "Size") - .map(name -> $x(String.format(сolumnHeaderLocator, name))) + .map(name -> $x(String.format(columnHeaderLocator, name))) .collect(Collectors.toList()); } private List<SelenideElement> getEnabledColumnHeaders(){ return Stream.of("Topic Name", "Partitions", "Out of sync replicas", "Size") - .map(name -> $x(String.format(сolumnHeaderLocator, name))) + .map(name -> $x(String.format(columnHeaderLocator, name))) .collect(Collectors.toList()); } diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WaitUtils.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WaitUtils.java deleted file mode 100644 index ec4ee223cb8..00000000000 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/utilities/WaitUtils.java +++ /dev/null @@ -1,31 +0,0 @@ -package com.provectus.kafka.ui.utilities; - -import com.codeborne.selenide.Condition; -import com.codeborne.selenide.SelenideElement; -import org.junit.jupiter.api.Assertions; -import org.openqa.selenium.By; - -import static com.codeborne.selenide.Selenide.*; -import static com.codeborne.selenide.Selenide.$; - -public class WaitUtils { - public static void refreshUntil(By by, Condition condition) { - int i = 0; - do { - refresh(); - i++; - sleep(2000); - } while ($$(by).size() < 1 && i != 20); - $(by).shouldBe(condition); - } - - public static void waitForSelectedValue(SelenideElement element, String selectedValue) { - int i = 0; - do { - refresh(); - i++; - sleep(2000); - } while (!selectedValue.equals(element.getSelectedValue()) && i != 60); - Assertions.assertEquals(selectedValue, element.getSelectedValue()) ; - } -}
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/schemas/SchemasTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/schemas/SchemasTests.java index 553568795dd..398130bcd19 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/schemas/SchemasTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/schemas/SchemasTests.java @@ -103,25 +103,52 @@ void updateSchemaAvro() { Assertions.assertEquals(CompatibilityLevel.CompatibilityEnum.NONE.toString(), schemaDetails.getCompatibility(), "getCompatibility()"); } + @DisplayName("Checking Compare Versions operation for Schema") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(186) + @Test + @Order(3) + void compareVersionsOperation() { + naviSideBar + .openSideMenu(SCHEMA_REGISTRY); + schemaRegistryList + .waitUntilScreenReady() + .openSchema(AVRO_API.getName()); + int latestVersion = schemaDetails + .waitUntilScreenReady() + .getLatestVersion(); + schemaDetails + .openCompareVersionMenu(); + int versionsNumberFromDdl = schemaCreateForm + .waitUntilScreenReady() + .openSchemaVersionDdl() + .getVersionsNumberFromList(); + Assertions.assertEquals(latestVersion,versionsNumberFromDdl,"Versions number is not matched"); + schemaCreateForm + .selectVersionFromDropDown(1); + Assertions.assertEquals(53, schemaCreateForm.getMarkedLinesNumber(), "getAllMarkedLines()"); + } + @DisplayName("should delete AVRO schema") @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) @AutomationStatus(status = Status.AUTOMATED) @CaseId(187) @Test - @Order(3) + @Order(4) void deleteSchemaAvro() { - naviSideBar - .openSideMenu(SCHEMA_REGISTRY); - schemaRegistryList - .waitUntilScreenReady() - .openSchema(AVRO_API.getName()); - schemaDetails - .waitUntilScreenReady() - .removeSchema(); - schemaRegistryList - .waitUntilScreenReady(); - Assertions.assertFalse(schemaRegistryList.isSchemaVisible(AVRO_API.getName()),"isSchemaVisible()"); - SCHEMA_LIST.remove(AVRO_API); + naviSideBar + .openSideMenu(SCHEMA_REGISTRY); + schemaRegistryList + .waitUntilScreenReady() + .openSchema(AVRO_API.getName()); + schemaDetails + .waitUntilScreenReady() + .removeSchema(); + schemaRegistryList + .waitUntilScreenReady(); + Assertions.assertFalse(schemaRegistryList.isSchemaVisible(AVRO_API.getName()),"isSchemaVisible()"); + SCHEMA_LIST.remove(AVRO_API); } @DisplayName("should create JSON schema") @@ -129,7 +156,7 @@ void deleteSchemaAvro() { @AutomationStatus(status = Status.AUTOMATED) @CaseId(89) @Test - @Order(4) + @Order(5) void createSchemaJson() { Schema schemaJson = Schema.createSchemaJson(); naviSideBar @@ -162,7 +189,7 @@ void createSchemaJson() { @AutomationStatus(status = Status.AUTOMATED) @CaseId(189) @Test - @Order(5) + @Order(6) void deleteSchemaJson() { naviSideBar .openSideMenu(SCHEMA_REGISTRY); @@ -183,7 +210,7 @@ void deleteSchemaJson() { @AutomationStatus(status = Status.AUTOMATED) @CaseId(91) @Test - @Order(6) + @Order(7) void createSchemaProtobuf() { Schema schemaProtobuf = Schema.createSchemaProtobuf(); naviSideBar @@ -216,7 +243,7 @@ void createSchemaProtobuf() { @AutomationStatus(status = Status.AUTOMATED) @CaseId(223) @Test - @Order(7) + @Order(8) void deleteSchemaProtobuf() { naviSideBar .openSideMenu(SCHEMA_REGISTRY);
train
val
2022-11-21T12:44:22
"2022-11-14T14:04:37Z"
anezboretskiy
train
provectus/kafka-ui/2802_2952
provectus/kafka-ui
provectus/kafka-ui/2802
provectus/kafka-ui/2952
[ "keyword_pr_to_issue" ]
7e82e7434ea2d555d881e03a230e5e21db0636d4
fdde380072f8a269607f6ac0ef683813f31a0141
[ "Hi, I love your site. Would you mind if I work on this? ", "Hey, sure. Please.Β " ]
[]
"2022-11-18T14:10:06Z"
[ "type/bug", "good first issue", "scope/frontend", "status/accepted", "status/confirmed" ]
Create connector button is unavailable until the 2nd click
1. Click create connector button 2. Enter valid json config 3. Switch to the name field, enter the name 4. The button is still unavailable until you click inside the config text area again
[ "kafka-ui-react-app/src/components/Connect/New/New.tsx" ]
[ "kafka-ui-react-app/src/components/Connect/New/New.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Connect/New/New.tsx b/kafka-ui-react-app/src/components/Connect/New/New.tsx index 2040f238bd2..2e93b4f5ce5 100644 --- a/kafka-ui-react-app/src/components/Connect/New/New.tsx +++ b/kafka-ui-react-app/src/components/Connect/New/New.tsx @@ -42,7 +42,7 @@ const New: React.FC = () => { const mutation = useCreateConnector(clusterName); const methods = useForm<FormValues>({ - mode: 'onTouched', + mode: 'all', resolver: yupResolver(validationSchema), defaultValues: { connectName: get(connects, '0.name', ''),
null
train
val
2022-11-22T18:08:52
"2022-10-21T20:43:33Z"
Haarolean
train
provectus/kafka-ui/2860_2970
provectus/kafka-ui
provectus/kafka-ui/2860
provectus/kafka-ui/2970
[ "keyword_pr_to_issue" ]
7e82e7434ea2d555d881e03a230e5e21db0636d4
1ba8d14819067ef1e109b73e92d3ba59fed4a7f9
[]
[]
"2022-11-23T16:27:20Z"
[ "good first issue", "scope/frontend", "status/accepted", "type/chore" ]
KSQL: Focus onto query field after incorrect query execution
**Set up** [ee920b4](https://github.com/provectus/kafka-ui/commit/ee920b4) **Steps to Reproduce** <!-- We'd like you to provide an example setup (via docker-compose, helm, etc.) to reproduce the problem, especially with a complex setups. --> Steps to reproduce the behavior: 1. Select KSQL in the side menu 2. Click "Execute KSQL request" 3. Enter incorrect KSQL statement 4. Click "Execute" **Actual behavior** Error message is displayed (which is correct), but the focus is not moved to the KSQL entry field **Expected behavior** It is better to change focus back to the KSQL field, as it is most likely, that problem needs to be fixed there (a typo or missed semicolon **Additional context** Focus is working ok, if the query is executed via keyboard (CMD+ENTER/CTRL+ENTER)
[ "kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/QueryForm.tsx" ]
[ "kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/QueryForm.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/QueryForm.tsx b/kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/QueryForm.tsx index 9965ea5189a..26622881f72 100644 --- a/kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/QueryForm.tsx +++ b/kafka-ui-react-app/src/components/KsqlDb/Query/QueryForm/QueryForm.tsx @@ -1,4 +1,4 @@ -import React, { useCallback } from 'react'; +import React, { useCallback, useRef } from 'react'; import { FormError } from 'components/common/Input/Input.styled'; import { ErrorMessage } from '@hookform/error-message'; import { useForm, Controller, useFieldArray } from 'react-hook-form'; @@ -8,6 +8,7 @@ import CloseIcon from 'components/common/Icons/CloseIcon'; import { yupResolver } from '@hookform/resolvers/yup'; import yup from 'lib/yupExtended'; import PlusIcon from 'components/common/Icons/PlusIcon'; +import ReactAce from 'react-ace/lib/ace'; import * as S from './QueryForm.styled'; @@ -75,6 +76,17 @@ const QueryForm: React.FC<Props> = ({ } }, []); + const inputRef = useRef<ReactAce>(null); + + const handleFocus = () => { + // eslint-disable-next-line @typescript-eslint/no-explicit-any + const textInput = inputRef?.current?.editor?.textInput as any; + + if (textInput) { + textInput.focus(); + } + }; + return ( <S.QueryWrapper> <form onSubmit={handleSubmit(submitHandler)}> @@ -111,6 +123,7 @@ const QueryForm: React.FC<Props> = ({ }, ]} readOnly={fetching} + ref={inputRef} /> )} /> @@ -188,6 +201,7 @@ const QueryForm: React.FC<Props> = ({ buttonSize="M" type="submit" disabled={fetching} + onClick={handleFocus} > Execute </Button>
null
test
val
2022-11-22T18:08:52
"2022-11-01T14:30:10Z"
BulatKha
train
provectus/kafka-ui/2965_2972
provectus/kafka-ui
provectus/kafka-ui/2965
provectus/kafka-ui/2972
[ "connected" ]
7fe999660c93e516a5e95729a358a2eafec5889e
5f232a3966cda70eac34aa2b32ec4a610404d1c6
[]
[ "are you sure this case will run before deleteTopic() every time without test run order specifying?", "here we don't need soft assertion, because it's a new step in test don't related to previous action. is there's any mismatch inside the getEnabledActionButtons() loop test should fail there", "why do we need this duplicated method? seems we can set getActionButtons public with Step annotation and use it instead", "1. method works with the one element -> it should have name of this element in the title\n2. seems better to use boolean (ex: isCopySelectedTopicBtnEnabled()) and assert it inside test as isTrue os isFalse", "this method will select and UNSELECT checkbox because u don't check its state inside. we hase isSelected() method in selenide \nu can do like if(!element.isSelected()) element.click() -> this way method can be named 'selectCheckboxByName'\nand the main point -> i've asked u to take different functions to avoid merge conflicts. Arthur now works on implementing the inner class to work with this grid. and your implementation is temporary here", "Done", "done", "Done", "Added @TestMethodOrder(MethodOrderer.OrderAnnotation.class) and sorted tests with order", "Refactored with Assertions.assertFalse", "can we use here and at 121 line some default instances? seems if topic creation by api will fail we'll lose atomicity and get additional failed test", "changed to connect_configs and connect_offset topics " ]
"2022-11-23T20:31:13Z"
[ "scope/QA", "scope/AQA" ]
Checking available operations for selected Topic within 'All Topics' page
Auto test implementation for: https://app.qase.io/case/KAFKAUI-7 Description: Checking available operations for selected Topic within 'All Topics' page Pre-conditions: Login to Kafka-ui application Open the 'Local' section Select the 'Topics' Steps: 1. Check any Topic from 'All Topics' list Expected result: Should display 'Delete selected topics', 'Copy selected topic' 'Purge messages of selected topics' options 2. Check one more Topic from 'All Topics' list Expected result: Should disappear 'Copy selected topic' function
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java index c32621bf1a5..11d47a07ca3 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java @@ -18,7 +18,10 @@ public class TopicsList extends BasePage { protected SelenideElement addTopicBtn = $x("//button[normalize-space(text()) ='Add a Topic']"); protected SelenideElement searchField = $x("//input[@placeholder='Search by Topic Name']"); protected SelenideElement showInternalRadioBtn = $x("//input[@name='ShowInternalTopics']"); - protected String actionButtonLocator = "//button[text()='%s']"; + protected SelenideElement deleteSelectedTopicsBtn = $x("//button[text()='Delete selected topics']"); + protected SelenideElement copySelectedTopicBtn = $x("//button[text()='Copy selected topic']"); + protected SelenideElement purgeMessagesOfSelectedTopicsBtn = $x("//button[text()='Purge messages of selected topics']"); + protected String checkBoxListLocator = "//a[@title='%s']//ancestor::td/../td/input[@type='checkbox']"; @Step public TopicsList waitUntilScreenReady() { @@ -45,9 +48,21 @@ public TopicsList openTopic(String topicName) { return this; } - private List<SelenideElement> getActionButtons() { - return Stream.of("Delete selected topics", "Copy selected topic", "Purge messages of selected topics") - .map(name -> $x(String.format(actionButtonLocator, name))) + @Step + public TopicsList selectCheckboxByName(String topicName){ + SelenideElement checkBox = $x(String.format(checkBoxListLocator,topicName)); + if(!checkBox.is(Condition.selected)){clickByJavaScript(checkBox);} + return this; + } + + @Step + public boolean isCopySelectedTopicBtnEnabled(){ + return isEnabled(copySelectedTopicBtn); + } + + @Step + public List<SelenideElement> getActionButtons() { + return Stream.of(deleteSelectedTopicsBtn, copySelectedTopicBtn, purgeMessagesOfSelectedTopicsBtn) .collect(Collectors.toList()); }
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java index c342a216074..05b614c3bbb 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java @@ -27,10 +27,14 @@ import org.junit.jupiter.api.BeforeAll; import org.junit.jupiter.api.Disabled; import org.junit.jupiter.api.DisplayName; +import org.junit.jupiter.api.MethodOrderer; +import org.junit.jupiter.api.Order; import org.junit.jupiter.api.Test; import org.junit.jupiter.api.TestInstance; +import org.junit.jupiter.api.TestMethodOrder; @TestInstance(TestInstance.Lifecycle.PER_CLASS) +@TestMethodOrder(MethodOrderer.OrderAnnotation.class) public class TopicsTests extends BaseTest { private static final long SUITE_ID = 2; private static final String SUITE_TITLE = "Topics"; @@ -63,6 +67,7 @@ public void beforeAll() { @AutomationStatus(status = Status.AUTOMATED) @CaseId(199) @Test + @Order(1) public void createTopic() { naviSideBar .openSideMenu(TOPICS); @@ -97,6 +102,28 @@ public void createTopic() { TOPIC_LIST.add(TOPIC_TO_CREATE); } + @DisplayName("Checking available operations for selected Topic within 'All Topics' page") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(7) + @Test + @Order(2) + void checkAvailableOperations(){ + String processingTopic = "my_ksql_1ksql_processing_log"; + String confluentTopic = "_confluent-ksql-my_ksql_1_command_topic"; + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .selectCheckboxByName(processingTopic); + topicsList.getActionButtons(). + forEach(element -> assertThat(element.is(Condition.enabled)) + .as(element.getSearchCriteria() + " isEnabled()").isTrue()); + topicsList + .selectCheckboxByName(confluentTopic); + Assertions.assertFalse(topicsList.isCopySelectedTopicBtnEnabled(),"isCopySelectedTopicBtnEnabled()"); + } + @Disabled() @Issue("https://github.com/provectus/kafka-ui/issues/2625") @DisplayName("should update a topic") @@ -104,6 +131,7 @@ public void createTopic() { @AutomationStatus(status = Status.AUTOMATED) @CaseId(197) @Test + @Order(3) public void updateTopic() { naviSideBar .openSideMenu(TOPICS); @@ -150,6 +178,7 @@ public void updateTopic() { @AutomationStatus(status = Status.AUTOMATED) @CaseId(207) @Test + @Order(4) public void deleteTopic() { naviSideBar .openSideMenu(TOPICS); @@ -174,6 +203,7 @@ public void deleteTopic() { @AutomationStatus(status = Status.AUTOMATED) @CaseId(20) @Test + @Order(5) void redirectToConsumerFromTopic() { String topicName = "source-activities"; String consumerGroupId = "connect-sink_postgres_activities"; @@ -199,6 +229,7 @@ void redirectToConsumerFromTopic() { @AutomationStatus(status = Status.AUTOMATED) @CaseId(4) @Test + @Order(6) void checkTopicCreatePossibility() { naviSideBar .openSideMenu(TOPICS); @@ -219,6 +250,7 @@ void checkTopicCreatePossibility() { @AutomationStatus(status = Status.AUTOMATED) @CaseId(6) @Test + @Order(7) void checkCustomParametersWithinCreateNewTopic() { naviSideBar .openSideMenu(TOPICS); @@ -243,6 +275,7 @@ void checkCustomParametersWithinCreateNewTopic() { @AutomationStatus(status = Status.AUTOMATED) @CaseId(2) @Test + @Order(8) void checkTopicListElements() { naviSideBar .openSideMenu(TOPICS); @@ -263,6 +296,7 @@ void checkTopicListElements() { @AutomationStatus(status = Status.AUTOMATED) @CaseId(12) @Test + @Order(9) void addingNewFilterWithinTopic() { String topicName = "_schemas"; String filterName = "123ABC";
train
val
2022-11-25T15:23:30
"2022-11-22T11:47:49Z"
anezboretskiy
train
provectus/kafka-ui/2963_2986
provectus/kafka-ui
provectus/kafka-ui/2963
provectus/kafka-ui/2986
[ "keyword_pr_to_issue" ]
9b76d595132955e6048bb7f6b40a3a74bf53209f
c8619268cd1e791b5d509371981cb0d9b0a80e99
[]
[]
"2022-11-29T12:02:53Z"
[ "good first issue", "scope/frontend", "status/accepted", "type/chore" ]
Topics: make the topic name field focused
<!-- Don't forget to check for existing issues/discussions regarding your proposal. We might already have it. https://github.com/provectus/kafka-ui/issues https://github.com/provectus/kafka-ui/discussions --> **Describe the bug** <!--(A clear and concise description of what the bug is.)--> When a new topic is created via UI, it is easier for user to key in the new topic name and fill in the rest via Tab button from the keyboard. **Set up** <!-- How do you run the app? Please provide as much info as possible: 1. App version (docker image version or check commit hash in the top left corner in UI) 2. Helm chart version, if you use one 3. Any IAAC configs We might close the issue without further explanation if you don't provide such information. --> [6cf8fb3](https://github.com/provectus/kafka-ui/commit/6cf8fb3) **Steps to Reproduce** <!-- We'd like you to provide an example setup (via docker-compose, helm, etc.) to reproduce the problem, especially with a complex setups. --> Steps to reproduce the behavior: 1. Go to topics and click the "Add a topic" button. **Actual behavior** No edit field is in the focus, while the page only has them and no other specific fields, that could not be entered from the keyboard. **Expected behavior** <!-- (A clear and concise description of what you expected to happen) --> Topic name would be the first thing to enter on the Create topic page, and user should be able to type it straight in, without clicking on the field first. **Screenshots** <!-- (If applicable, add screenshots to help explain your problem) --> **Additional context** <!-- (Add any other context about the problem here) -->
[ "kafka-ui-react-app/src/components/Connect/New/New.tsx", "kafka-ui-react-app/src/components/Schemas/New/New.tsx", "kafka-ui-react-app/src/components/Topics/shared/Form/TopicForm.tsx" ]
[ "kafka-ui-react-app/src/components/Connect/New/New.tsx", "kafka-ui-react-app/src/components/Schemas/New/New.tsx", "kafka-ui-react-app/src/components/Topics/shared/Form/TopicForm.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Connect/New/New.tsx b/kafka-ui-react-app/src/components/Connect/New/New.tsx index 2e399359425..1d3f230631a 100644 --- a/kafka-ui-react-app/src/components/Connect/New/New.tsx +++ b/kafka-ui-react-app/src/components/Connect/New/New.tsx @@ -133,6 +133,7 @@ const New: React.FC = () => { inputSize="M" placeholder="Connector Name" name="name" + autoFocus autoComplete="off" disabled={isSubmitting} /> diff --git a/kafka-ui-react-app/src/components/Schemas/New/New.tsx b/kafka-ui-react-app/src/components/Schemas/New/New.tsx index 0d6e9d8ffa5..3e5d39c9df5 100644 --- a/kafka-ui-react-app/src/components/Schemas/New/New.tsx +++ b/kafka-ui-react-app/src/components/Schemas/New/New.tsx @@ -103,6 +103,7 @@ const New: React.FC = () => { <Input inputSize="M" placeholder="Schema Name" + autoFocus name="subject" autoComplete="off" disabled={isSubmitting} diff --git a/kafka-ui-react-app/src/components/Topics/shared/Form/TopicForm.tsx b/kafka-ui-react-app/src/components/Topics/shared/Form/TopicForm.tsx index 15f60eaaac2..ae9d429236a 100644 --- a/kafka-ui-react-app/src/components/Topics/shared/Form/TopicForm.tsx +++ b/kafka-ui-react-app/src/components/Topics/shared/Form/TopicForm.tsx @@ -93,6 +93,7 @@ const TopicForm: React.FC<Props> = ({ <InputLabel htmlFor="topicFormName">Topic Name *</InputLabel> <Input id="topicFormName" + autoFocus name="name" placeholder="Topic Name" defaultValue={topicName}
null
test
val
2023-02-27T09:50:12
"2022-11-21T19:28:57Z"
BulatKha
train
provectus/kafka-ui/2982_2992
provectus/kafka-ui
provectus/kafka-ui/2982
provectus/kafka-ui/2992
[ "connected" ]
27db304b1b3f779363891dccdaf365e17dbaf593
86c72d7d8527e00aaa5ac45a04a8909e7a6a21fb
[]
[ "remove System.out.println pls", "use custom line breaks pls in cases where line is too long", "can we use Condition.enabled here instead of handling attributes?", "Done", "Done", "Refactored" ]
"2022-11-30T09:11:46Z"
[ "scope/QA", "scope/AQA" ]
[e2e]SchemasTests.updateSchemaAvro : Update schema avro ( REFACTORED WITH NEW ASSERTIONS)
Auto test implementation for: https://app.qase.io/case/KAFKAUI-186 Description: The purpose of this case is to make sure that possible to edit the Schemas Pre-conditions: - Login to Kafka-ui - Navigate to Schema Registry - Create a schema - Open the Schema Steps: - [ ] 1. Press "Edit Schema" button Expected result Following content should be displayed: - Disabled Type field - Compatibility level - Latest Schema - New schema - Disabled "Submit" button - [x] 2. Make change within schema body Expected result the "Submit" button should become active - [x] 3. Press "Compatibility level" dropdown field Input data Change any value Expected result The available values should be listed - [x] 4. Select any value for "Compatibility level" Input data "NONE" - [x] 5. Click on "Submit" button Expected result The Schema profile should open - [x] 6. Press "Edit Schema" button Expected result All changes should be saved
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaCreateForm.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaCreateForm.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/schemas/SchemasTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaCreateForm.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaCreateForm.java index 7d0d5c58a93..d749b5feb8b 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaCreateForm.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/schema/SchemaCreateForm.java @@ -11,6 +11,8 @@ import com.provectus.kafka.ui.pages.BasePage; import io.qameta.allure.Step; import java.util.List; +import java.util.stream.Collectors; +import java.util.stream.Stream; public class SchemaCreateForm extends BasePage { @@ -21,6 +23,7 @@ public class SchemaCreateForm extends BasePage { protected SelenideElement schemaTypeDdl = $x("//ul[@name='schemaType']"); protected SelenideElement compatibilityLevelList = $x("//ul[@name='compatibilityLevel']"); protected SelenideElement newSchemaTextArea = $x("//div[@id='newSchema']"); + protected SelenideElement latestSchemaTextArea = $x("//div[@id='latestSchema']"); protected SelenideElement schemaVersionDdl = $$x("//ul[@role='listbox']/li[text()='Version 2']").first(); protected List<SelenideElement> visibleMarkers = $$x("//div[@class='ace_scroller']//div[contains(@class,'codeMarker')]"); protected List<SelenideElement> elementsCompareVersionDdl = $$x("//ul[@role='listbox']/ul/li"); @@ -96,14 +99,25 @@ public SchemaCreateForm setNewSchemaValue(String configJson) { } @Step - public boolean isSchemaDropDownDisabled(){ - boolean disabled = false; + public List<SelenideElement> getAllDetailsPageElements() { + return Stream.of(compatibilityLevelList, newSchemaTextArea, latestSchemaTextArea, submitBtn, schemaTypeDdl) + .collect(Collectors.toList()); + } + + @Step + public boolean isSubmitBtnEnabled(){ + return isEnabled(submitBtn); + } + + @Step + public boolean isSchemaDropDownEnabled(){ + boolean enabled = true; try{ String attribute = schemaTypeDdl.getAttribute("disabled"); - disabled = true; + enabled = false; } catch (Throwable ignored){ } - return disabled; + return enabled; } }
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/schemas/SchemasTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/schemas/SchemasTests.java index 398130bcd19..f112c7b5881 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/schemas/SchemasTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/schemas/SchemasTests.java @@ -4,6 +4,7 @@ import static com.provectus.kafka.ui.settings.Source.CLUSTER_NAME; import static com.provectus.kafka.ui.utilities.FileUtils.fileToString; +import com.codeborne.selenide.Condition; import com.provectus.kafka.ui.api.model.CompatibilityLevel; import com.provectus.kafka.ui.base.BaseTest; import com.provectus.kafka.ui.models.Schema; @@ -93,7 +94,13 @@ void updateSchemaAvro() { .openEditSchema(); schemaCreateForm .waitUntilScreenReady(); - Assertions.assertTrue(schemaCreateForm.isSchemaDropDownDisabled(),"isSchemaDropDownDisabled()"); + SoftAssertions softly = new SoftAssertions(); + schemaCreateForm.getAllDetailsPageElements() + .forEach(element -> softly.assertThat(element.is(Condition.visible)) + .as(element.getSearchCriteria() + " isVisible").isTrue()); + softly.assertThat(schemaCreateForm.isSubmitBtnEnabled()).as("isSubmitBtnEnabled()").isFalse(); + softly.assertThat(schemaCreateForm.isSchemaDropDownEnabled()).as("isSchemaDropDownEnabled()").isFalse(); + softly.assertAll(); schemaCreateForm .selectCompatibilityLevelFromDropdown(CompatibilityLevel.CompatibilityEnum.NONE) .setNewSchemaValue(fileToString(AVRO_API.getValuePath()))
val
val
2022-11-29T16:15:08
"2022-11-28T13:34:00Z"
anezboretskiy
train
provectus/kafka-ui/2968_3001
provectus/kafka-ui
provectus/kafka-ui/2968
provectus/kafka-ui/3001
[ "connected" ]
c804a6d00bce6eee1108740ba7714ebf7bc6668d
052d8af45632265cc88192663b14fec5aff713f5
[ "Hello there iblis17! πŸ‘‹\n\nThank you and congratulations πŸŽ‰ for opening your very first issue in this project! πŸ’–\n\nIn case you want to claim this issue, please comment down below! We will try to get back to you as soon as we can. πŸ‘€", "Hey, thanks for reaching out. We'll take a look!" ]
[]
"2022-12-01T11:52:57Z"
[ "type/bug", "scope/backend", "status/accepted" ]
Deleting a "custom parameter" of topic doesn't take effect
<!-- Don't forget to check for existing issues/discussions regarding your proposal. We might already have it. https://github.com/provectus/kafka-ui/issues https://github.com/provectus/kafka-ui/discussions --> **Describe the bug** <!--(A clear and concise description of what the bug is.)--> Deleting a "custom parameter" of topic doesn't reflect deletion to the broker at the page `/ui/clusters/local/all-topics/mytopic/edit`. **Steps to Reproduce** <!-- We'd like you to provide an example setup (via docker-compose, helm, etc.) to reproduce the problem, especially with a complex setups. --> I have kafka v3.3.1 cluster with 3 broker nodes in kraft mode. Steps to reproduce the behavior: 1. I use the docker image `provectuslabs/kafka-ui:7e82e7434ea2d555d881e03a230e5e21db0636d4` 2. Go to the topic edit page, e.g. `/ui/clusters/local/all-topics/mytopic/edit`. 3. Add a custom parameter, `compression.type` and set the value to `gzip`. 4. Go to the topic settings page and verify the modification (`/ui/clusters/local/all-topics/mytopic/settings`) 5. Go to the topic edit page and delete the custom parameter, then click "Update topic" button. 6. Go to the topic settings page but found the parameter not changed back to default value (`producer`). **Expected behavior** <!-- (A clear and concise description of what you expected to happen) --> In step6, the parameter `compression.type` should be set to the default value `producer`. I expected the deleting a custom parameter mimics the behaviour of ``` $ bin/kafka-configs.sh --bootstrap-server localhost:9092 --entity-type topics --entity-name mytopic --alter --delete-config compression.type ```
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java" ]
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java" ]
[ "kafka-ui-api/src/test/java/com/provectus/kafka/ui/service/ReactiveAdminClientTest.java" ]
diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java index 2504473b171..ebaa8475595 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java @@ -32,6 +32,8 @@ import java.util.stream.Collectors; import java.util.stream.Stream; import javax.annotation.Nullable; +import lombok.AccessLevel; +import lombok.Getter; import lombok.RequiredArgsConstructor; import lombok.Value; import lombok.extern.slf4j.Slf4j; @@ -147,6 +149,7 @@ private static <T> Mono<T> toMono(KafkaFuture<T> future) { //--------------------------------------------------------------------------------- + @Getter(AccessLevel.PACKAGE) // visible for testing private final AdminClient client; private final String version; private final Set<SupportedFeature> features; @@ -361,9 +364,14 @@ public Mono<Void> createPartitions(Map<String, NewPartitions> newPartitionsMap) return toMono(client.createPartitions(newPartitionsMap).all()); } + + // NOTE: places whole current topic config with new one. Entries that were present in old config, + // but missed in new will be set to default public Mono<Void> updateTopicConfig(String topicName, Map<String, String> configs) { if (features.contains(SupportedFeature.INCREMENTAL_ALTER_CONFIGS)) { - return incrementalAlterConfig(topicName, configs); + return getTopicsConfigImpl(List.of(topicName), false) + .map(conf -> conf.getOrDefault(topicName, List.of())) + .flatMap(currentConfigs -> incrementalAlterConfig(topicName, currentConfigs, configs)); } else { return alterConfig(topicName, configs); } @@ -499,17 +507,22 @@ public Mono<Void> alterReplicaLogDirs(Map<TopicPartitionReplica, String> replica return toMono(client.alterReplicaLogDirs(replicaAssignment).all()); } - private Mono<Void> incrementalAlterConfig(String topicName, Map<String, String> configs) { - var config = configs.entrySet().stream() - .flatMap(cfg -> Stream.of( - new AlterConfigOp( - new ConfigEntry( - cfg.getKey(), - cfg.getValue()), - AlterConfigOp.OpType.SET))) - .collect(toList()); - var topicResource = new ConfigResource(ConfigResource.Type.TOPIC, topicName); - return toMono(client.incrementalAlterConfigs(Map.of(topicResource, config)).all()); + private Mono<Void> incrementalAlterConfig(String topicName, + List<ConfigEntry> currentConfigs, + Map<String, String> newConfigs) { + var configsToDelete = currentConfigs.stream() + .filter(e -> e.source() == ConfigEntry.ConfigSource.DYNAMIC_TOPIC_CONFIG) //manually set configs only + .filter(e -> !newConfigs.containsKey(e.name())) + .map(e -> new AlterConfigOp(e, AlterConfigOp.OpType.DELETE)); + + var configsToSet = newConfigs.entrySet().stream() + .map(e -> new AlterConfigOp(new ConfigEntry(e.getKey(), e.getValue()), AlterConfigOp.OpType.SET)); + + return toMono(client.incrementalAlterConfigs( + Map.of( + new ConfigResource(ConfigResource.Type.TOPIC, topicName), + Stream.concat(configsToDelete, configsToSet).toList() + )).all()); } @SuppressWarnings("deprecation")
diff --git a/kafka-ui-api/src/test/java/com/provectus/kafka/ui/service/ReactiveAdminClientTest.java b/kafka-ui-api/src/test/java/com/provectus/kafka/ui/service/ReactiveAdminClientTest.java new file mode 100644 index 00000000000..a84b42a4e3f --- /dev/null +++ b/kafka-ui-api/src/test/java/com/provectus/kafka/ui/service/ReactiveAdminClientTest.java @@ -0,0 +1,92 @@ +package com.provectus.kafka.ui.service; + +import static java.util.Objects.requireNonNull; +import static org.assertj.core.api.Assertions.assertThat; + +import com.provectus.kafka.ui.AbstractIntegrationTest; +import java.util.ArrayList; +import java.util.List; +import java.util.Map; +import java.util.UUID; +import java.util.stream.Stream; +import lombok.SneakyThrows; +import org.apache.kafka.clients.admin.AdminClient; +import org.apache.kafka.clients.admin.AlterConfigOp; +import org.apache.kafka.clients.admin.Config; +import org.apache.kafka.clients.admin.ConfigEntry; +import org.apache.kafka.clients.admin.NewTopic; +import org.apache.kafka.common.config.ConfigResource; +import org.junit.function.ThrowingRunnable; +import org.junit.jupiter.api.AfterEach; +import org.junit.jupiter.api.BeforeEach; +import org.junit.jupiter.api.Test; +import reactor.test.StepVerifier; + +class ReactiveAdminClientTest extends AbstractIntegrationTest { + + private final List<ThrowingRunnable> clearings = new ArrayList<>(); + + private AdminClient adminClient; + private ReactiveAdminClient reactiveAdminClient; + + @BeforeEach + void init() { + AdminClientService adminClientService = applicationContext.getBean(AdminClientService.class); + ClustersStorage clustersStorage = applicationContext.getBean(ClustersStorage.class); + reactiveAdminClient = requireNonNull(adminClientService.get(clustersStorage.getClusterByName(LOCAL).get()).block()); + adminClient = reactiveAdminClient.getClient(); + } + + @AfterEach + void tearDown() { + for (ThrowingRunnable clearing : clearings) { + try { + clearing.run(); + } catch (Throwable th) { + //NOOP + } + } + } + + @Test + void testUpdateTopicConfigs() throws Exception { + String topic = UUID.randomUUID().toString(); + createTopics(new NewTopic(topic, 1, (short) 1)); + + var configResource = new ConfigResource(ConfigResource.Type.TOPIC, topic); + + adminClient.incrementalAlterConfigs( + Map.of( + configResource, + List.of( + new AlterConfigOp(new ConfigEntry("compression.type", "gzip"), AlterConfigOp.OpType.SET), + new AlterConfigOp(new ConfigEntry("retention.bytes", "12345678"), AlterConfigOp.OpType.SET) + ) + ) + ).all().get(); + + StepVerifier.create( + reactiveAdminClient.updateTopicConfig( + topic, + Map.of( + "compression.type", "snappy", //changing existing config + "file.delete.delay.ms", "12345" // adding new one + ) + ) + ).expectComplete().verify(); + + Config config = adminClient.describeConfigs(List.of(configResource)).values().get(configResource).get(); + assertThat(config.get("retention.bytes").value()).isNotEqualTo("12345678"); // wes reset to default + assertThat(config.get("compression.type").value()).isEqualTo("snappy"); + assertThat(config.get("file.delete.delay.ms").value()).isEqualTo("12345"); + } + + + @SneakyThrows + void createTopics(NewTopic... topics) { + adminClient.createTopics(List.of(topics)).all().get(); + clearings.add(() -> adminClient.deleteTopics(Stream.of(topics).map(NewTopic::name).toList()).all().get()); + } + + +}
train
val
2022-12-01T10:20:48
"2022-11-23T08:03:01Z"
iblislin
train
provectus/kafka-ui/2995_3003
provectus/kafka-ui
provectus/kafka-ui/2995
provectus/kafka-ui/3003
[ "connected" ]
86c72d7d8527e00aaa5ac45a04a8909e7a6a21fb
a503c29472bbfc8f4c57680fd71ba7bfb007d6f9
[]
[ "don't we need to wait until new screen ready?", "seems all that variables we use only once as argument for some methods\nsuggest to not create extra instances in this case\nrefactor your test and all another tests inside this class", "here u're already at topicDetails, no reason in 359 line", "why do we need this assert?", "seems it's not a locator of the grid", "setSaveThisFilterCheckbox -> selectSaveThisFilterCheckbox\nu setting the boolean as argument and it's hard to understand what means set(true) in checkbox case", "isSavedFilterVisible - where? at the filters list of inside the AddFilterMdl?", "at the mdl?", "Added waiter", "Done", "refactored", "my mistake , deleted ", "change to savedFiltersNameLocator", "done ", "change to isSavedFilterVisibleAtFiltersList", "refactored\r\n", "I've asked u to remove only those variables that are used only once", "update this assertion with new created method isSavedFilterVisibleAtFiltersList, please\nold instances u can remove", "Revert this local variables and once more as String filterName = \"123ABC\" in addingNewFilterWithinTopic()", "Done", "is that instance at modal? mb we can call it setDisplayNameFldAddFilterMdl for better understanding?", "selectSaveThisFilterCheckbox -> selectSaveThisFilterCheckboxMdl", "savedFiltersNameLocator -> savedFilterNameLocator", ".as(\"isFilterNameVisible(filterName)\")?", "Done", "Done", "done", "refactored to ()" ]
"2022-12-01T14:23:34Z"
[ "scope/QA", "scope/AQA" ]
[e2e]Checking filter saving within Messages/Topic profile/Saved Filters
Auto test implementation for: https://app.qase.io/case/KAFKAUI-13 Description: Checking filter saving within Messages/Topic profile Pre-conditions: - Login to Kafka-ui application - Open the 'Local' section - Select the 'Topics' - Open the Topic profile - Turn to 'Messages' tab Steps: 1. Press '+Add Filters' button Expected result: 'Add Filter' pop up should open 2. Fill fields Expected result: Make sure 'Add filter' button becomes active 3. Check 'Save this filter' checkbox 4. Click on 'Saved Filters' link Expected result: New created filter should display within saved filters list
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java index 95c3e6fb23f..d6fd9af61d3 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java @@ -23,7 +23,7 @@ public class TopicDetails extends BasePage { protected SelenideElement overviewTab = $x("//a[contains(text(),'Overview')]"); protected SelenideElement messagesTab = $x("//a[contains(text(),'Messages')]"); protected SelenideElement addFiltersBtn = $x("//button[text()='Add Filters']"); - protected SelenideElement savedFiltersField = $x("//div[text()='Saved Filters']"); + protected SelenideElement savedFiltersLink = $x("//div[text()='Saved Filters']"); protected SelenideElement addFilterCodeModalTitle = $x("//label[text()='Filter code']"); protected SelenideElement addFilterCodeInput = $x("//div[@id='ace-editor']//textarea"); protected SelenideElement saveThisFilterCheckBoxAddFilterMdl = $x("//input[@name='saveFilter']"); @@ -37,10 +37,12 @@ public class TopicDetails extends BasePage { protected SelenideElement contentMessageTab = $x("//html//div[@id='root']/div/main//table//p"); protected SelenideElement cleanUpPolicyField = $x("//div[contains(text(),'Clean Up Policy')]/../span/*"); protected SelenideElement partitionsField = $x("//div[contains(text(),'Partitions')]/../span"); + protected SelenideElement backToCreateFiltersLink = $x("//div[text()='Back To create filters']"); protected ElementsCollection messageGridItems = $$x("//tbody//tr"); + protected String savedFilterNameLocator = "//div[@role='savedFilter']/div[contains(text(),'%s')]"; protected String consumerIdLocator = "//a[@title='%s']"; protected String topicHeaderLocator = "//h1[contains(text(),'%s')]"; - protected String filterNameLocator = "//*[@data-testid='activeSmartFilter']"; + protected String activeFilterNameLocator = "//div[@data-testid='activeSmartFilter'][contains(text(),'%s')]"; @Step public TopicDetails waitUntilScreenReady() { @@ -120,15 +122,20 @@ public TopicDetails clickMessagesAddFiltersBtn() { } @Step - public TopicDetails waitUntilAddFiltersMdlVisible() { - addFilterCodeModalTitle.shouldBe(Condition.visible); + public TopicDetails openSavedFiltersListMdl(){ + savedFiltersLink.shouldBe(Condition.enabled).click(); + backToCreateFiltersLink.shouldBe(Condition.visible); return this; } @Step - public TopicDetails clickAddFilterBtnAddFilterMdl() { - addFilterBtnAddFilterMdl.shouldBe(Condition.enabled).click(); - addFilterCodeModalTitle.shouldBe(Condition.hidden); + public boolean isFilterVisibleAtSavedFiltersMdl(String filterName){ + return isVisible($x(String.format(savedFilterNameLocator,filterName))); + } + + @Step + public TopicDetails waitUntilAddFiltersMdlVisible() { + addFilterCodeModalTitle.shouldBe(Condition.visible); return this; } @@ -138,23 +145,46 @@ public TopicDetails setFilterCodeFieldAddFilterMdl(String filterCode) { return this; } + @Step + public TopicDetails selectSaveThisFilterCheckboxMdl(boolean select){ + selectElement(saveThisFilterCheckBoxAddFilterMdl, select); + return this; + } + @Step public boolean isSaveThisFilterCheckBoxSelected() { return isSelected(saveThisFilterCheckBoxAddFilterMdl); } + @Step + public TopicDetails setDisplayNameFldAddFilterMdl(String displayName) { + displayNameInputAddFilterMdl.shouldBe(Condition.enabled).sendKeys(displayName); + return this; + } + + @Step + public TopicDetails clickAddFilterBtnAndCloseMdl(boolean closeModal) { + addFilterBtnAddFilterMdl.shouldBe(Condition.enabled).click(); + if(closeModal){ + addFilterCodeModalTitle.shouldBe(Condition.hidden);} + else{ + addFilterCodeModalTitle.shouldBe(Condition.visible); + } + return this; + } + @Step public boolean isAddFilterBtnAddFilterMdlEnabled() { return isEnabled(addFilterBtnAddFilterMdl); } @Step - public String getFilterName() { - return $x(filterNameLocator).getText(); + public boolean isActiveFilterVisible(String activeFilterName) { + return isVisible($x(String.format(activeFilterNameLocator, activeFilterName))); } public List<SelenideElement> getAllAddFilterModalVisibleElements() { - return Arrays.asList(savedFiltersField, displayNameInputAddFilterMdl, addFilterBtnAddFilterMdl, cancelBtnAddFilterMdl); + return Arrays.asList(savedFiltersLink, displayNameInputAddFilterMdl, addFilterBtnAddFilterMdl, cancelBtnAddFilterMdl); } public List<SelenideElement> getAllAddFilterModalEnabledElements() {
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java index 51b182bb413..a90b299cf25 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java @@ -1,6 +1,7 @@ package com.provectus.kafka.ui.suite.topics; import static com.provectus.kafka.ui.pages.NaviSideBar.SideMenuOption.TOPICS; +import static com.provectus.kafka.ui.pages.topic.TopicDetails.TopicMenu.MESSAGES; import static com.provectus.kafka.ui.pages.topic.enums.CleanupPolicyValue.COMPACT; import static com.provectus.kafka.ui.pages.topic.enums.CleanupPolicyValue.DELETE; import static com.provectus.kafka.ui.pages.topic.enums.CustomParameterType.COMPRESSION_TYPE; @@ -8,6 +9,7 @@ import static com.provectus.kafka.ui.settings.Source.CLUSTER_NAME; import static com.provectus.kafka.ui.utilities.FileUtils.fileToString; import static org.apache.commons.lang.RandomStringUtils.randomAlphabetic; +import static org.apache.commons.lang.RandomStringUtils.randomAlphanumeric; import static org.assertj.core.api.Assertions.assertThat; import com.codeborne.selenide.Condition; @@ -109,20 +111,18 @@ public void createTopic() { @Test @Order(2) void checkAvailableOperations() { - String processingTopic = "my_ksql_1ksql_processing_log"; - String confluentTopic = "_confluent-ksql-my_ksql_1_command_topic"; naviSideBar .openSideMenu(TOPICS); topicsList .waitUntilScreenReady() - .getTopicItem(processingTopic) + .getTopicItem("my_ksql_1ksql_processing_log") .selectItem(true); topicsList .getActionButtons() .forEach(element -> assertThat(element.is(Condition.enabled)) .as(element.getSearchCriteria() + " isEnabled()").isTrue()); topicsList - .getTopicItem(confluentTopic) + .getTopicItem("_confluent-ksql-my_ksql_1_command_topic") .selectItem(true); Assertions.assertFalse(topicsList.isCopySelectedTopicBtnEnabled(), "isCopySelectedTopicBtnEnabled()"); } @@ -301,15 +301,15 @@ void checkTopicListElements() { @Test @Order(9) void addingNewFilterWithinTopic() { - String topicName = "_schemas"; - String filterName = "123ABC"; + String filterName = randomAlphanumeric(5); naviSideBar .openSideMenu(TOPICS); topicsList .waitUntilScreenReady() - .openTopic(topicName); + .openTopic("_schemas"); topicDetails - .openDetailsTab(TopicDetails.TopicMenu.MESSAGES) + .waitUntilScreenReady() + .openDetailsTab(MESSAGES) .clickMessagesAddFiltersBtn() .waitUntilAddFiltersMdlVisible(); SoftAssertions softly = new SoftAssertions(); @@ -327,11 +327,41 @@ void addingNewFilterWithinTopic() { softly.assertAll(); topicDetails .setFilterCodeFieldAddFilterMdl(filterName); - assertThat(topicDetails.isAddFilterBtnAddFilterMdlEnabled()).as("isMessagesAddFilterTabAddFilterBtnEnabled()") + assertThat(topicDetails.isAddFilterBtnAddFilterMdlEnabled()).as("isAddFilterBtnAddFilterMdlEnabled()") + .isTrue(); + topicDetails.clickAddFilterBtnAndCloseMdl(true); + assertThat(topicDetails.isActiveFilterVisible(filterName)).as("isActiveFilterVisible()") .isTrue(); - topicDetails.clickAddFilterBtnAddFilterMdl(); - assertThat(topicDetails.getFilterName()).as("isFilterNameVisible(filterName)") - .isEqualTo(filterName); + } + + @DisplayName("Checking filter saving within Messages/Topic profile/Saved Filters") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(13) + @Test + @Order(10) + void checkFilterSavingWithinSavedFilters() { + String displayName = randomAlphanumeric(5); + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .openTopic("my_ksql_1ksql_processing_log"); + topicDetails + .waitUntilScreenReady() + .openDetailsTab(MESSAGES) + .clickMessagesAddFiltersBtn() + .waitUntilAddFiltersMdlVisible() + .setFilterCodeFieldAddFilterMdl(randomAlphanumeric(4)) + .selectSaveThisFilterCheckboxMdl(true) + .setDisplayNameFldAddFilterMdl(displayName); + assertThat(topicDetails.isAddFilterBtnAddFilterMdlEnabled()).as("isAddFilterBtnAddFilterMdlEnabled()") + .isTrue(); + topicDetails + .clickAddFilterBtnAndCloseMdl(false) + .openSavedFiltersListMdl(); + assertThat(topicDetails.isFilterVisibleAtSavedFiltersMdl(displayName)) + .as("isFilterVisibleAtSavedFiltersMdl()").isTrue(); } @DisplayName("Checking 'Show Internal Topics' toggle functionality within 'All Topics' page") @@ -339,7 +369,7 @@ void addingNewFilterWithinTopic() { @AutomationStatus(status = Status.AUTOMATED) @CaseId(11) @Test - @Order(10) + @Order(11) void checkShowInternalTopicsButtonFunctionality(){ naviSideBar .openSideMenu(TOPICS);
val
val
2022-11-30T15:01:09
"2022-11-30T13:50:56Z"
anezboretskiy
train
provectus/kafka-ui/1252_3004
provectus/kafka-ui
provectus/kafka-ui/1252
provectus/kafka-ui/3004
[ "connected" ]
8717416d85d81275541444841e2764ca168af9fb
43ccca43c2c1aa2cbb6ed9fc97ef99e17daa63bb
[ "Need to investigate.", "This is what I get when reproducing it **java.util.concurrent.CompletionException: org.apache.kafka.common.errors.UnknownTopicOrPartitionException: This server does not host this topic-partition.**\r\n\r\nIt does randomly happen.\r\n\r\nhttps://issues.apache.org/jira/browse/KAFKA-6221", "Blocked by kafka's issue" ]
[]
"2022-12-01T17:57:08Z"
[ "type/bug", "good first issue", "scope/backend", "status/accepted", "status/blocked" ]
Redesign: topic creation fails randomly
Sometimes upon creating a topic this happens: `404 Topic not found` Clicking the button again produces this: `400 Topic 'deltest' already exists.`.
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java" ]
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java" ]
[ "kafka-ui-api/src/test/java/com/provectus/kafka/ui/service/ReactiveAdminClientTest.java" ]
diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java index ebaa8475595..c655cc2ad9b 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java @@ -3,6 +3,7 @@ import static com.google.common.util.concurrent.Uninterruptibles.getUninterruptibly; import static java.util.stream.Collectors.toList; import static java.util.stream.Collectors.toMap; +import static org.apache.kafka.clients.admin.ListOffsetsResult.ListOffsetsResultInfo; import com.google.common.collect.ImmutableMap; import com.google.common.collect.Iterators; @@ -16,6 +17,7 @@ import java.util.ArrayList; import java.util.Arrays; import java.util.Collection; +import java.util.HashMap; import java.util.HashSet; import java.util.Iterator; import java.util.List; @@ -45,6 +47,7 @@ import org.apache.kafka.clients.admin.ConsumerGroupListing; import org.apache.kafka.clients.admin.DescribeConfigsOptions; import org.apache.kafka.clients.admin.ListConsumerGroupOffsetsOptions; +import org.apache.kafka.clients.admin.ListOffsetsResult; import org.apache.kafka.clients.admin.ListTopicsOptions; import org.apache.kafka.clients.admin.NewPartitionReassignment; import org.apache.kafka.clients.admin.NewPartitions; @@ -170,6 +173,7 @@ public Mono<Map<String, List<ConfigEntry>>> getTopicsConfig() { return listTopics(true).flatMap(topics -> getTopicsConfig(topics, false)); } + //NOTE: skips not-found topics (for which UnknownTopicOrPartitionException was thrown by AdminClient) public Mono<Map<String, List<ConfigEntry>>> getTopicsConfig(Collection<String> topicNames, boolean includeDoc) { var includeDocFixed = features.contains(SupportedFeature.CONFIG_DOCUMENTATION_RETRIEVAL) && includeDoc; // we need to partition calls, because it can lead to AdminClient timeouts in case of large topics count @@ -258,7 +262,7 @@ public Mono<TopicDescription> describeTopic(String topic) { * This method converts input map into Mono[Map] ignoring keys for which KafkaFutures * finished with <code>clazz</code> exception. */ - private <K, V> Mono<Map<K, V>> toMonoWithExceptionFilter(Map<K, KafkaFuture<V>> values, + static <K, V> Mono<Map<K, V>> toMonoWithExceptionFilter(Map<K, KafkaFuture<V>> values, Class<? extends KafkaException> clazz) { if (values.isEmpty()) { return Mono.just(Map.of()); @@ -468,19 +472,26 @@ private Set<TopicPartition> filterPartitionsWithLeaderCheck(Collection<TopicDesc } // 1. NOTE(!): should only apply for partitions with existing leader, - // otherwise AdminClient will try to fetch topic metadata, fail and retry infinitely (until timeout) - // 2. TODO: check if it is a bug that AdminClient never throws LeaderNotAvailableException and just retrying instead + // otherwise AdminClient will try to fetch topic metadata, fail and retry infinitely (until timeout) + // 2. NOTE(!): Skips partitions that were not initialized yet + // (UnknownTopicOrPartitionException thrown, ex. after topic creation) + // 3. TODO: check if it is a bug that AdminClient never throws LeaderNotAvailableException and just retrying instead @KafkaClientInternalsDependant public Mono<Map<TopicPartition, Long>> listOffsetsUnsafe(Collection<TopicPartition> partitions, OffsetSpec offsetSpec) { Function<Collection<TopicPartition>, Mono<Map<TopicPartition, Long>>> call = - parts -> toMono( - client.listOffsets(parts.stream().collect(toMap(tp -> tp, tp -> offsetSpec))).all()) - .map(offsets -> offsets.entrySet().stream() - // filtering partitions for which offsets were not found - .filter(e -> e.getValue().offset() >= 0) - .collect(toMap(Map.Entry::getKey, e -> e.getValue().offset()))); + parts -> { + ListOffsetsResult r = client.listOffsets(parts.stream().collect(toMap(tp -> tp, tp -> offsetSpec))); + Map<TopicPartition, KafkaFuture<ListOffsetsResultInfo>> perPartitionResults = new HashMap<>(); + parts.forEach(p -> perPartitionResults.put(p, r.partitionResult(p))); + + return toMonoWithExceptionFilter(perPartitionResults, UnknownTopicOrPartitionException.class) + .map(offsets -> offsets.entrySet().stream() + // filtering partitions for which offsets were not found + .filter(e -> e.getValue().offset() >= 0) + .collect(toMap(Map.Entry::getKey, e -> e.getValue().offset()))); + }; return partitionCalls( partitions,
diff --git a/kafka-ui-api/src/test/java/com/provectus/kafka/ui/service/ReactiveAdminClientTest.java b/kafka-ui-api/src/test/java/com/provectus/kafka/ui/service/ReactiveAdminClientTest.java index a84b42a4e3f..26df31426da 100644 --- a/kafka-ui-api/src/test/java/com/provectus/kafka/ui/service/ReactiveAdminClientTest.java +++ b/kafka-ui-api/src/test/java/com/provectus/kafka/ui/service/ReactiveAdminClientTest.java @@ -1,9 +1,11 @@ package com.provectus.kafka.ui.service; +import static com.provectus.kafka.ui.service.ReactiveAdminClient.toMonoWithExceptionFilter; import static java.util.Objects.requireNonNull; import static org.assertj.core.api.Assertions.assertThat; import com.provectus.kafka.ui.AbstractIntegrationTest; +import com.provectus.kafka.ui.producer.KafkaTestProducer; import java.util.ArrayList; import java.util.List; import java.util.Map; @@ -15,7 +17,13 @@ import org.apache.kafka.clients.admin.Config; import org.apache.kafka.clients.admin.ConfigEntry; import org.apache.kafka.clients.admin.NewTopic; +import org.apache.kafka.clients.admin.OffsetSpec; +import org.apache.kafka.clients.producer.ProducerRecord; +import org.apache.kafka.common.KafkaFuture; +import org.apache.kafka.common.TopicPartition; import org.apache.kafka.common.config.ConfigResource; +import org.apache.kafka.common.errors.UnknownTopicOrPartitionException; +import org.apache.kafka.common.internals.KafkaFutureImpl; import org.junit.function.ThrowingRunnable; import org.junit.jupiter.api.AfterEach; import org.junit.jupiter.api.BeforeEach; @@ -88,5 +96,53 @@ void createTopics(NewTopic... topics) { clearings.add(() -> adminClient.deleteTopics(Stream.of(topics).map(NewTopic::name).toList()).all().get()); } + @Test + void testToMonoWithExceptionFilter() { + var failedFuture = new KafkaFutureImpl<String>(); + failedFuture.completeExceptionally(new UnknownTopicOrPartitionException()); + + var okFuture = new KafkaFutureImpl<String>(); + okFuture.complete("done"); + + Map<String, KafkaFuture<String>> arg = Map.of("failure", failedFuture, "ok", okFuture); + StepVerifier.create(toMonoWithExceptionFilter(arg, UnknownTopicOrPartitionException.class)) + .assertNext(result -> assertThat(result).hasSize(1).containsEntry("ok", "done")) + .verifyComplete(); + } + + @Test + void testListOffsetsUnsafe() { + String topic = UUID.randomUUID().toString(); + createTopics(new NewTopic(topic, 2, (short) 1)); + + // sending messages to have non-zero offsets for tp + try (var producer = KafkaTestProducer.forKafka(kafka)) { + producer.send(new ProducerRecord<>(topic, 1, "k", "v")); + producer.send(new ProducerRecord<>(topic, 1, "k", "v")); + } + + var requestedPartitions = List.of( + new TopicPartition(topic, 0), + new TopicPartition(topic, 1) + ); + + StepVerifier.create(reactiveAdminClient.listOffsetsUnsafe(requestedPartitions, OffsetSpec.earliest())) + .assertNext(offsets -> { + assertThat(offsets) + .hasSize(2) + .containsEntry(new TopicPartition(topic, 0), 0L) + .containsEntry(new TopicPartition(topic, 1), 0L); + }) + .verifyComplete(); + + StepVerifier.create(reactiveAdminClient.listOffsetsUnsafe(requestedPartitions, OffsetSpec.latest())) + .assertNext(offsets -> { + assertThat(offsets) + .hasSize(2) + .containsEntry(new TopicPartition(topic, 0), 0L) + .containsEntry(new TopicPartition(topic, 1), 2L); + }) + .verifyComplete(); + } }
val
val
2022-12-02T16:44:26
"2021-12-14T12:00:29Z"
Haarolean
train
provectus/kafka-ui/2907_3005
provectus/kafka-ui
provectus/kafka-ui/2907
provectus/kafka-ui/3005
[ "keyword_pr_to_issue" ]
c8619268cd1e791b5d509371981cb0d9b0a80e99
e261143bb466998302e7b91ccd6ad07622bc548a
[]
[]
"2022-12-01T18:27:09Z"
[ "type/bug", "scope/frontend", "status/accepted", "status/confirmed" ]
[UI] Topics: Message tab "Oldest first" button overlaps the "Submit" button on pages narrower than 1100px
<!-- Don't forget to check for existing issues/discussions regarding your proposal. We might already have it. https://github.com/provectus/kafka-ui/issues https://github.com/provectus/kafka-ui/discussions --> **Describe the bug** <!--(A clear and concise description of what the bug is.)--> **Set up** <!-- How do you run the app? Please provide as much info as possible: 1. App version (docker image version or check commit hash in the top left corner in UI) 2. Helm chart version, if you use one 3. Any IAAC configs We might close the issue without further explanation if you don't provide such information. --> [8011f60](https://github.com/provectus/kafka-ui/commit/8011f60) **Steps to Reproduce** <!-- We'd like you to provide an example setup (via docker-compose, helm, etc.) to reproduce the problem, especially with a complex setups. --> Steps to reproduce the behavior: 1. Open any topic 2. Switch to Messages tab and play with the browser window width **Actual behavior** - Message sorting drop-down list ("Oldest first", "Newest first" or "Live mode" overlays the controls to the left. And since it has transparent layout, looks very strange - Also the drop-down list values are not fully sown on the screen - Horizontal scroll is not available **Expected behavior** <!-- (A clear and concise description of what you expected to happen) --> Messages sorting drop-down list can be shown lower, where Search and "Add filters" button is located. **Screenshots** ![image](https://user-images.githubusercontent.com/112083452/200556537-60ae28ac-2638-492b-a53d-ae66a3819238.png) <!-- (If applicable, add screenshots to help explain your problem) --> **Additional context** <!-- (Add any other context about the problem here) --> There are related tasks: [UI] Timestamp value overlaps with dropdown field icon within Topic/Messages [#2858](https://github.com/provectus/kafka-ui/issues/2858) Improve the Responsive view for Messages/Topic page [#2303](https://github.com/provectus/kafka-ui/issues/2303)
[ "kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.styled.ts" ]
[ "kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.styled.ts" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.styled.ts b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.styled.ts index a8df901be4d..10802ee0db1 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.styled.ts +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.styled.ts @@ -33,6 +33,7 @@ export const FilterInputs = styled.div` gap: 8px; align-items: flex-end; width: 90%; + flex-wrap: wrap; `; export const SeekTypeSelectorWrapper = styled.div`
null
test
val
2023-02-27T11:44:28
"2022-11-08T11:59:45Z"
BulatKha
train
provectus/kafka-ui/2987_3014
provectus/kafka-ui
provectus/kafka-ui/2987
provectus/kafka-ui/3014
[ "connected" ]
a503c29472bbfc8f4c57680fd71ba7bfb007d6f9
0a24b2d7658a2fd9cfe5cb7af6961edb2c8fe1f6
[]
[ "why do we need order annotation is this class?\nseems we don't mind which test will run first, because their instances aren't depended", "seems we use this check at SmokeTest, no reason to duplicate it here. it's not correspond to main goal of this test", "getBroker -> GetAllBrokers\nno need to filter by visibility, because another way table won't be created", "method name should start from lower case", "getVisibleUptimeSummaryCells -> getUptimeSummaryCells", "getVisiblePartitionsSummaryCells -> getPartitionsSummaryCells", "kinky method. you trying to assert the broker's id for already shown broker in table? suggest remove it", "Removed.", "Renamened.", "Renamened.", "Fixed.", "Removed.", "Removed." ]
"2022-12-02T12:18:18Z"
[ "scope/QA", "scope/AQA" ]
[e2e]Checking the Brokers overview
Autotest implementation for: https://app.qase.io/case/KAFKAUI-1 Description: Checking the existing brokers in a cluster Pre-conditions: Empty value Post-conditions: Empty value Steps: 1 Login to Kafka-ui 2 Press the cluster in a left panel 3 Press the Brokers. Expected result: 1 The existing clusters should be displayed in a left 2 Brokers, Topics, Consumers, Schema Registry, Kafka connect, KSQL DB components should be visible 3 More than 0 Brokers should be available with following information: Uptime, Partitions, Broker ID, Segment Size, Segment Count, Port, Host
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersList.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersList.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java index 069b1b41dda..d7b342b0254 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java @@ -1,8 +1,10 @@ package com.provectus.kafka.ui.pages; +import static com.codeborne.selenide.Selenide.$$x; import static com.codeborne.selenide.Selenide.$x; import com.codeborne.selenide.Condition; +import com.codeborne.selenide.ElementsCollection; import com.codeborne.selenide.SelenideElement; import com.provectus.kafka.ui.utilities.WebUtils; import lombok.extern.slf4j.Slf4j; @@ -16,6 +18,7 @@ public abstract class BasePage extends WebUtils { protected SelenideElement dotMenuBtn = $x("//button[@aria-label='Dropdown Toggle']"); protected SelenideElement alertHeader = $x("//div[@role='alert']//div[@role='heading']"); protected SelenideElement alertMessage = $x("//div[@role='alert']//div[@role='contentinfo']"); + protected ElementsCollection allGridItems = $$x("//tr[@class]"); protected String summaryCellLocator = "//div[contains(text(),'%s')]"; protected String tableElementNameLocator = "//tbody//a[contains(text(),'%s')]"; protected String columnHeaderLocator = "//table//tr/th/div[text()='%s']"; diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersList.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersList.java index 0d9ee67492c..3df1a09304b 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersList.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersList.java @@ -2,10 +2,15 @@ import static com.codeborne.selenide.Selenide.$x; +import com.codeborne.selenide.CollectionCondition; import com.codeborne.selenide.Condition; import com.codeborne.selenide.SelenideElement; import com.provectus.kafka.ui.pages.BasePage; import io.qameta.allure.Step; +import java.util.ArrayList; +import java.util.List; +import java.util.stream.Collectors; +import java.util.stream.Stream; public class BrokersList extends BasePage { @@ -19,14 +24,100 @@ public BrokersList waitUntilScreenReady() { } @Step - public boolean isBrokerVisible(String brokerId) { - tableGrid.shouldBe(Condition.visible); - return isVisible(getTableElement(brokerId)); + public BrokersList openBroker(String brokerId) { + getBrokerItem(brokerId).openItem(); + return this; + } + + private List<SelenideElement> getUptimeSummaryCells() { + return Stream.of("Broker Count", "Active Controllers", "Version") + .map(name -> $x(String.format(summaryCellLocator, name))) + .collect(Collectors.toList()); + } + + private List<SelenideElement> getPartitionsSummaryCells() { + return Stream.of("Online", "URP", "In Sync Replicas", "Out Of Sync Replicas") + .map(name -> $x(String.format(summaryCellLocator, name))) + .collect(Collectors.toList()); } @Step - public BrokersList openBroker(String brokerName) { - getTableElement(brokerName).shouldBe(Condition.enabled).click(); - return this; + public List<SelenideElement> getAllVisibleElements() { + List<SelenideElement> visibleElements = new ArrayList<>(getUptimeSummaryCells()); + visibleElements.addAll(getPartitionsSummaryCells()); + return visibleElements; + } + + private List<SelenideElement> getEnabledColumnHeaders() { + return Stream.of("Broker ID", "Segment Size", "Segment Count", "Port", "Host") + .map(name -> $x(String.format(columnHeaderLocator, name))) + .collect(Collectors.toList()); + } + + @Step + public List<SelenideElement> getAllEnabledElements() { + return getEnabledColumnHeaders(); + } + + private List<BrokersList.BrokerGridItem> initGridItems() { + List<BrokersList.BrokerGridItem> gridItemList = new ArrayList<>(); + allGridItems.shouldHave(CollectionCondition.sizeGreaterThan(0)) + .forEach(item -> gridItemList.add(new BrokersList.BrokerGridItem(item))); + return gridItemList; + } + + @Step + public BrokerGridItem getBrokerItem(String id){ + return initGridItems().stream() + .filter(e ->e.getId().equals(id)) + .findFirst().orElse(null); + } + + @Step + public List<BrokerGridItem> getAllBrokers(){ + return initGridItems(); + } + + public static class BrokerGridItem extends BasePage { + + private final SelenideElement element; + + public BrokerGridItem(SelenideElement element) { + this.element = element; + } + + private SelenideElement getIdElm() { + return element.$x("./td[1]/a"); + } + + @Step + public String getId() { + return getIdElm().getText().trim(); + } + + @Step + public void openItem() { + getIdElm().click(); + } + + @Step + public int getSegmentSize(){ + return Integer.parseInt(element.$x("./td[2]").getText().trim()); + } + + @Step + public int getSegmentCount(){ + return Integer.parseInt(element.$x("./td[3]").getText().trim()); + } + + @Step + public int getPort(){ + return Integer.parseInt(element.$x("./td[4]").getText().trim()); + } + + @Step + public String getHost(){ + return element.$x("./td[5]").getText().trim(); + } } } diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java index d6fd9af61d3..9c6c4cec936 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java @@ -218,7 +218,7 @@ public int getMessageCountAmount() { private List<TopicDetails.MessageGridItem> initItems() { List<TopicDetails.MessageGridItem> gridItemList = new ArrayList<>(); - messageGridItems.shouldHave(CollectionCondition.sizeGreaterThan(0)) + allGridItems.shouldHave(CollectionCondition.sizeGreaterThan(0)) .forEach(item -> gridItemList.add(new TopicDetails.MessageGridItem(item))); return gridItemList; } diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java index 4c88f45aa90..871ecbb752f 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicsList.java @@ -1,11 +1,9 @@ package com.provectus.kafka.ui.pages.topic; -import static com.codeborne.selenide.Selenide.$$x; import static com.codeborne.selenide.Selenide.$x; import com.codeborne.selenide.CollectionCondition; import com.codeborne.selenide.Condition; -import com.codeborne.selenide.ElementsCollection; import com.codeborne.selenide.SelenideElement; import com.provectus.kafka.ui.pages.BasePage; import io.qameta.allure.Step; @@ -24,7 +22,6 @@ public class TopicsList extends BasePage { protected SelenideElement deleteSelectedTopicsBtn = $x("//button[text()='Delete selected topics']"); protected SelenideElement copySelectedTopicBtn = $x("//button[text()='Copy selected topic']"); protected SelenideElement purgeMessagesOfSelectedTopicsBtn = $x("//button[text()='Purge messages of selected topics']"); - protected ElementsCollection topicsGridItems = $$x("//tr[@class]"); @Step public TopicsList waitUntilScreenReady() { @@ -102,7 +99,7 @@ public List<SelenideElement> getAllEnabledElements() { private List<TopicGridItem> initGridItems() { List<TopicGridItem> gridItemList = new ArrayList<>(); - topicsGridItems.shouldHave(CollectionCondition.sizeGreaterThan(0)) + allGridItems.shouldHave(CollectionCondition.sizeGreaterThan(0)) .forEach(item -> gridItemList.add(new TopicGridItem(item))); return gridItemList; }
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java index 5a493438f98..dcff92bc28c 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java @@ -17,6 +17,25 @@ public class BrokersTests extends BaseTest { private static final String SUITE_TITLE = "Brokers"; private static final long SUITE_ID = 1; + @DisplayName("Checking the Brokers overview") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(1) + @Test + public void checkBrokersOverview(){ + naviSideBar + .openSideMenu(BROKERS); + brokersList + .waitUntilScreenReady(); + SoftAssertions softly = new SoftAssertions(); + softly.assertThat(brokersList.getAllBrokers()).as("getAllBrokers()").size().isGreaterThan(0); + brokersList.getAllVisibleElements().forEach(element -> softly.assertThat(element.is(Condition.visible)) + .as(element.getSearchCriteria() + " isVisible()").isTrue()); + brokersList.getAllEnabledElements().forEach(element -> softly.assertThat(element.is(Condition.enabled)) + .as(element.getSearchCriteria() + " isEnabled()").isTrue()); + softly.assertAll(); + } + @DisplayName("Checking the existing Broker's profile in a cluster") @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) @AutomationStatus(status = Status.AUTOMATED) @@ -27,7 +46,7 @@ public void checkExistingBrokersInCluster(){ .openSideMenu(BROKERS); brokersList .waitUntilScreenReady(); - assertThat(brokersList.isBrokerVisible("1")).as("isBrokerVisible()").isTrue(); + assertThat(brokersList.getAllBrokers()).as("getAllBrokers()").size().isGreaterThan(0); brokersList .openBroker("1"); brokersDetails
train
val
2022-12-02T15:30:10
"2022-11-29T13:09:43Z"
ArthurNiedial
train
provectus/kafka-ui/3016_3023
provectus/kafka-ui
provectus/kafka-ui/3016
provectus/kafka-ui/3023
[ "connected" ]
25d34e11e4a6e6b472a2a60d3c103be444a319eb
d2d039e36c7e78343e4b0d98632a891982660695
[]
[ "not sure we really need this linebreak", "I believe we need to wait until screen ready first", "TOPIC_TO_RETAIN_DATA rename to cameCase pls\nfor example -> topicWithTimeToRetain", "let's keep a single style in namings. all topics in this class are named using next pattern: topic-name-randomAlphabetic(5)", "Deleted ", "added ", "Yes , it*s not a constant variable , refactored ", "changed to randomAlphabetic(5)" ]
"2022-12-05T14:16:41Z"
[ "scope/QA", "scope/AQA" ]
[e2e]Checking "Time to retain data (in ms)" custom value with editing Topic's settings
Autotest implementation for: https://app.qase.io/case/KAFKAUI-266 Description: Purpose of this case is to check custom "Time to retain data (in ms)" to not be default value always Pre-conditions: Login to https://www.kafka-ui.provectus.io/ Post-conditions: Empty value Steps: 1. Navigate to Topics from left navigation bar 2. Add a Topic Expected result: Topic form should open with "604800000" default value for "Time to retain data (in ms)" 3. Fill Topic name 4. Change "Time to retain data (in ms)" value Input data: "86400000" 1 day 5. Press "Create Topic" Expected result: New created Topic profile should open 6. Select "Edit settings" from menu icon Expected result: Custom "86400000" value for "Time to retain data (in ms)" should be displayed instead of default value
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java index 9c6c4cec936..eb14687ed99 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java @@ -43,6 +43,7 @@ public class TopicDetails extends BasePage { protected String consumerIdLocator = "//a[@title='%s']"; protected String topicHeaderLocator = "//h1[contains(text(),'%s')]"; protected String activeFilterNameLocator = "//div[@data-testid='activeSmartFilter'][contains(text(),'%s')]"; + protected String settingsGridValueLocator = "//tbody/tr/td/span[text()='%s']//ancestor::tr/td[2]/span"; @Step public TopicDetails waitUntilScreenReady() { @@ -58,6 +59,11 @@ public TopicDetails openDetailsTab(TopicMenu menu) { return this; } + @Step + public String getSettingsGridValueByKey(String key){ + return $x(String.format(settingsGridValueLocator, key)).scrollTo().shouldBe(Condition.visible).getText(); + } + @Step public TopicDetails openDotMenu() { clickByJavaScript(dotMenuBtn);
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java index a90b299cf25..559baaecc63 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java @@ -2,6 +2,7 @@ import static com.provectus.kafka.ui.pages.NaviSideBar.SideMenuOption.TOPICS; import static com.provectus.kafka.ui.pages.topic.TopicDetails.TopicMenu.MESSAGES; +import static com.provectus.kafka.ui.pages.topic.TopicDetails.TopicMenu.SETTINGS; import static com.provectus.kafka.ui.pages.topic.enums.CleanupPolicyValue.COMPACT; import static com.provectus.kafka.ui.pages.topic.enums.CleanupPolicyValue.DELETE; import static com.provectus.kafka.ui.pages.topic.enums.CustomParameterType.COMPRESSION_TYPE; @@ -248,12 +249,47 @@ void checkTopicCreatePossibility() { assertThat(topicCreateEditForm.isCreateTopicButtonEnabled()).as("isCreateTopicButtonEnabled()").isTrue(); } + @DisplayName("Checking 'Time to retain data (in ms)' custom value with editing Topic's settings") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(266) + @Test + @Order(7) + void checkTimeToRetainDataCustomValueWithEditingTopic() { + Topic topicToRetainData = new Topic() + .setName("topic-to-retain-data-" + randomAlphabetic(5)) + .setTimeToRetainData("86400000"); + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .clickAddTopicBtn(); + topicCreateEditForm + .waitUntilScreenReady() + .setTopicName(topicToRetainData.getName()); + assertThat(topicCreateEditForm.getTimeToRetain()).as("getTimeToRetain()").isEqualTo("604800000"); + topicCreateEditForm + .setTimeToRetainDataInMs(topicToRetainData.getTimeToRetainData()) + .clickCreateTopicBtn(); + topicDetails + .waitUntilScreenReady() + .openDotMenu() + .clickEditSettingsMenu(); + assertThat(topicCreateEditForm.getTimeToRetain()).as("getTimeToRetain()") + .isEqualTo(topicToRetainData.getTimeToRetainData()); + topicDetails + .openDetailsTab(SETTINGS); + assertThat(topicDetails.getSettingsGridValueByKey("retention.ms")).as("getSettingsGridValueByKey()") + .isEqualTo(topicToRetainData.getTimeToRetainData()); + TOPIC_LIST.add(topicToRetainData); + } + @DisplayName("Checking requiredness of Custom parameters within 'Create new Topic'") @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) @AutomationStatus(status = Status.AUTOMATED) @CaseId(6) @Test - @Order(7) + @Order(8) void checkCustomParametersWithinCreateNewTopic() { naviSideBar .openSideMenu(TOPICS); @@ -278,7 +314,7 @@ void checkCustomParametersWithinCreateNewTopic() { @AutomationStatus(status = Status.AUTOMATED) @CaseId(2) @Test - @Order(8) + @Order(9) void checkTopicListElements() { naviSideBar .openSideMenu(TOPICS); @@ -299,9 +335,9 @@ void checkTopicListElements() { @AutomationStatus(status = Status.AUTOMATED) @CaseId(12) @Test - @Order(9) + @Order(10) void addingNewFilterWithinTopic() { - String filterName = randomAlphanumeric(5); + String filterName = randomAlphabetic(5); naviSideBar .openSideMenu(TOPICS); topicsList @@ -339,9 +375,9 @@ void addingNewFilterWithinTopic() { @AutomationStatus(status = Status.AUTOMATED) @CaseId(13) @Test - @Order(10) + @Order(11) void checkFilterSavingWithinSavedFilters() { - String displayName = randomAlphanumeric(5); + String displayName = randomAlphabetic(5); naviSideBar .openSideMenu(TOPICS); topicsList @@ -369,7 +405,7 @@ void checkFilterSavingWithinSavedFilters() { @AutomationStatus(status = Status.AUTOMATED) @CaseId(11) @Test - @Order(11) + @Order(12) void checkShowInternalTopicsButtonFunctionality(){ naviSideBar .openSideMenu(TOPICS);
val
val
2022-12-06T05:52:29
"2022-12-02T15:04:33Z"
anezboretskiy
train
provectus/kafka-ui/3013_3028
provectus/kafka-ui
provectus/kafka-ui/3013
provectus/kafka-ui/3028
[ "connected" ]
5d31189609ddb7024d5f16a4f9ed52cbb9022355
fdf8db98a22c7403c2910b326dbe6458d3ddcbba
[ "The problem is not only with the `$` , any bit complex jsonPath does not work with it , cause it uses `Lodash` `get` method, instead of an actually JsonPath identifier, hence other stuff will fail as well.\r\n\r\nExample `reddit-comments-avro` from topics messages check the `$.map.array[1:2]`, `map.array[1:2]` stuff like this will not work cause lodash does not support it.\r\n\r\nExpressions such as\r\n\r\n* `[?(expression)]`\r\n* `[-n:]`\r\n\r\n\r\nMethod will itself will return undefined over that particular JSON hence the bug.\r\n\r\nOne of the solution is to integrate a JSON PATH library , but those tend to be larger in size hence it will increase our download time of javascript bundle.\r\n\r\n\r\n\r\n<br class=\"Apple-interchange-newline\">\r\n\r\n", "@armenuikafka please double-check the test case has this case.", "> @armenuikafka please double-check the test case has this case.\r\n\r\n@BulatKha test case for this issue - https://app.qase.io/case/KAFKAUI-270" ]
[]
"2022-12-06T12:00:55Z"
[ "scope/frontend", "status/accepted", "type/chore" ]
Fields previews: make the paths work as json paths
currently they work without `$`
[ "kafka-ui-react-app/package.json", "kafka-ui-react-app/pnpm-lock.yaml", "kafka-ui-react-app/src/components/Topics/Topic/Messages/Message.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/PreviewModal.styled.ts" ]
[ "kafka-ui-react-app/package.json", "kafka-ui-react-app/pnpm-lock.yaml", "kafka-ui-react-app/src/components/Topics/Topic/Messages/Message.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/PreviewModal.styled.ts" ]
[]
diff --git a/kafka-ui-react-app/package.json b/kafka-ui-react-app/package.json index 5c51e700076..94759221dfd 100644 --- a/kafka-ui-react-app/package.json +++ b/kafka-ui-react-app/package.json @@ -26,6 +26,7 @@ "jest": "^29.0.3", "jest-watch-typeahead": "^2.0.0", "json-schema-faker": "^0.5.0-rcv.44", + "jsonpath-plus": "^7.2.0", "lodash": "^4.17.21", "pretty-ms": "7.0.1", "react": "^18.1.0", diff --git a/kafka-ui-react-app/pnpm-lock.yaml b/kafka-ui-react-app/pnpm-lock.yaml index 52390df0cde..eb9b03c0447 100644 --- a/kafka-ui-react-app/pnpm-lock.yaml +++ b/kafka-ui-react-app/pnpm-lock.yaml @@ -63,6 +63,7 @@ specifiers: jest-styled-components: ^7.0.8 jest-watch-typeahead: ^2.0.0 json-schema-faker: ^0.5.0-rcv.44 + jsonpath-plus: ^7.2.0 lint-staged: ^13.0.2 lodash: ^4.17.21 prettier: ^2.3.1 @@ -116,6 +117,7 @@ dependencies: jest: 29.0.3_yqiaopbgmqcuvx27p5xxvum6wm jest-watch-typeahead: [email protected] json-schema-faker: 0.5.0-rcv.44 + jsonpath-plus: 7.2.0 lodash: 4.17.21 pretty-ms: 7.0.1 react: 18.1.0 @@ -7058,6 +7060,11 @@ packages: engines: {node: '>=10.0.0'} dev: false + /jsonpath-plus/7.2.0: + resolution: {integrity: sha512-zBfiUPM5nD0YZSBT/o/fbCUlCcepMIdP0CJZxM1+KgA4f2T206f6VAg9e7mX35+KlMaIc5qXW34f3BnwJ3w+RA==} + engines: {node: '>=12.0.0'} + dev: false + /jsx-ast-utils/3.3.0: resolution: {integrity: sha512-XzO9luP6L0xkxwhIJMTJQpZo/eeN60K08jHdexfD569AGxeNug6UketeHXEhROoM8aR7EcUoOQmIhcJQjcuq8Q==} engines: {node: '>=4.0'} diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Message.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Message.tsx index 82a705c3a33..cb2dbbb3c0c 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Message.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Message.tsx @@ -1,4 +1,3 @@ -import get from 'lodash/get'; import React from 'react'; import styled from 'styled-components'; import useDataSaver from 'lib/hooks/useDataSaver'; @@ -7,6 +6,7 @@ import MessageToggleIcon from 'components/common/Icons/MessageToggleIcon'; import IconButtonWrapper from 'components/common/Icons/IconButtonWrapper'; import { Dropdown, DropdownItem } from 'components/common/Dropdown'; import { formatTimestamp } from 'lib/dateTimeHelpers'; +import { JSONPath } from 'jsonpath-plus'; import MessageContent from './MessageContent/MessageContent'; import * as S from './MessageContent/MessageContent.styled'; @@ -88,9 +88,12 @@ const Message: React.FC<Props> = ({ return ( <> {filters.map((item) => ( - <div> - {item.field}: {get(parsedJson, item.path)} - </div> + <span key={`${item.path}--${item.field}`}> + {item.field}:{' '} + {JSON.stringify( + JSONPath({ path: item.path, json: parsedJson, wrap: false }) + )} + </span> ))} </> ); @@ -116,7 +119,7 @@ const Message: React.FC<Props> = ({ <StyledDataCell title={key}> {renderFilteredJson(key, keyFilters)} </StyledDataCell> - <StyledDataCell> + <StyledDataCell title={content}> <S.Metadata> <S.MetadataValue> {renderFilteredJson(content, contentFilters)} diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx index 96e9f40d73d..10e4f4dfa17 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessagesTable.tsx @@ -52,7 +52,18 @@ const MessagesTable: React.FC = () => { }; return ( - <> + <div style={{ position: 'relative' }}> + {previewFor !== null && ( + <PreviewModal + values={previewFor === 'key' ? keyFilters : contentFilters} + toggleIsOpen={() => setPreviewFor(null)} + setFilters={(payload: PreviewFilter[]) => + previewFor === 'key' + ? setKeyFilters(payload) + : setContentFilters(payload) + } + /> + )} <Table isFullwidth> <thead> <tr> @@ -77,18 +88,6 @@ const MessagesTable: React.FC = () => { onPreview={() => setPreviewFor('content')} /> <TableHeaderCell> </TableHeaderCell> - - {previewFor !== null && ( - <PreviewModal - values={previewFor === 'key' ? keyFilters : contentFilters} - toggleIsOpen={() => setPreviewFor(null)} - setFilters={(payload: PreviewFilter[]) => - previewFor === 'key' - ? setKeyFilters(payload) - : setContentFilters(payload) - } - /> - )} </tr> </thead> <tbody> @@ -139,7 +138,7 @@ const MessagesTable: React.FC = () => { </Button> </S.Pages> </S.Pagination> - </> + </div> ); }; diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/PreviewModal.styled.ts b/kafka-ui-react-app/src/components/Topics/Topic/Messages/PreviewModal.styled.ts index 35f221ddeab..c1cecfbbd05 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/PreviewModal.styled.ts +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/PreviewModal.styled.ts @@ -7,6 +7,7 @@ export const PreviewModal = styled.div` background: ${({ theme }) => theme.modal.backgroundColor}; position: absolute; left: 25%; + top: 30px; // some margin border: 1px solid ${({ theme }) => theme.modal.border.contrast}; box-shadow: ${({ theme }) => theme.modal.shadow}; padding: 32px;
null
train
val
2023-01-17T19:38:44
"2022-12-02T11:01:01Z"
Haarolean
train
provectus/kafka-ui/3027_3029
provectus/kafka-ui
provectus/kafka-ui/3027
provectus/kafka-ui/3029
[ "connected" ]
d2d039e36c7e78343e4b0d98632a891982660695
b6b1b8bda58e1cf9a3478ccbdbcd7f3ec0b89102
[]
[ "can't see step according to this assert in test case", "can't see step according to this assert in test case", "seems here should be disappeared addFilterCodeModalTitle instead of selectFilterBtnAddFilterMdl", "wrong description", "deleted", "deleted", "refactored", "here u're already at topicDetails screen, no reason in 423 line", "Yes , sorry it*s my carelessness, refactored! " ]
"2022-12-06T12:59:04Z"
[ "scope/QA", "scope/AQA" ]
[e2e]Checking applying saved filter within Topic/Messages
Autotest implementation for: https://app.qase.io/case/KAFKAUI-14 Description: Checking applying saved filter within Messages/Topic profile Pre-conditions: - Login to Kafka-ui application - Open the 'Local' section - Select the 'Topics' - Open the Topic profile - Turn to Messages tab Post-conditions: Empty value Steps: 1. Press '+Add Filters' button Expected result: 'Add Filter' pop up opens 2. Click the 'Saved Filters' 3. Select the filter 4. Press 'Select Filter' button Expected result: - 'Add filter' pop up should disappear - New filter should appear near '+Add Filters' button - Corresponding data should display
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java index eb14687ed99..675857dfd7e 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java @@ -1,20 +1,18 @@ package com.provectus.kafka.ui.pages.topic; -import static com.codeborne.selenide.Selenide.$; -import static com.codeborne.selenide.Selenide.$$x; -import static com.codeborne.selenide.Selenide.$x; -import static org.apache.commons.lang.math.RandomUtils.nextInt; - import com.codeborne.selenide.CollectionCondition; import com.codeborne.selenide.Condition; import com.codeborne.selenide.ElementsCollection; import com.codeborne.selenide.SelenideElement; import com.provectus.kafka.ui.pages.BasePage; import io.qameta.allure.Step; +import org.openqa.selenium.By; import java.util.ArrayList; import java.util.Arrays; import java.util.List; -import org.openqa.selenium.By; + +import static com.codeborne.selenide.Selenide.*; +import static org.apache.commons.lang.math.RandomUtils.nextInt; public class TopicDetails extends BasePage { @@ -30,6 +28,7 @@ public class TopicDetails extends BasePage { protected SelenideElement displayNameInputAddFilterMdl = $x("//input[@placeholder='Enter Name']"); protected SelenideElement cancelBtnAddFilterMdl = $x("//button[text()='Cancel']"); protected SelenideElement addFilterBtnAddFilterMdl = $x("//button[text()='Add filter']"); + protected SelenideElement selectFilterBtnAddFilterMdl = $x("//button[text()='Select filter']"); protected SelenideElement editSettingsMenu = $x("//li[@role][contains(text(),'Edit settings')]"); protected SelenideElement removeTopicBtn = $x("//ul[@role='menu']//div[contains(text(),'Remove Topic')]"); protected SelenideElement confirmBtn = $x("//div[@role='dialog']//button[contains(text(),'Confirm')]"); @@ -139,6 +138,19 @@ public boolean isFilterVisibleAtSavedFiltersMdl(String filterName){ return isVisible($x(String.format(savedFilterNameLocator,filterName))); } + @Step + public TopicDetails selectFilterAtSavedFiltersMdl(String filterName){ + $x(String.format(savedFilterNameLocator, filterName)).shouldBe(Condition.enabled).click(); + return this; + } + + @Step + public TopicDetails clickSelectFilterBtnAtSavedFiltersMdl(){ + selectFilterBtnAddFilterMdl.shouldBe(Condition.enabled).click(); + addFilterCodeModalTitle.shouldBe(Condition.disappear); + return this; + } + @Step public TopicDetails waitUntilAddFiltersMdlVisible() { addFilterCodeModalTitle.shouldBe(Condition.visible);
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java index 559baaecc63..4f7c566fba4 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java @@ -10,7 +10,6 @@ import static com.provectus.kafka.ui.settings.Source.CLUSTER_NAME; import static com.provectus.kafka.ui.utilities.FileUtils.fileToString; import static org.apache.commons.lang.RandomStringUtils.randomAlphabetic; -import static org.apache.commons.lang.RandomStringUtils.randomAlphanumeric; import static org.assertj.core.api.Assertions.assertThat; import com.codeborne.selenide.Condition; @@ -388,7 +387,7 @@ void checkFilterSavingWithinSavedFilters() { .openDetailsTab(MESSAGES) .clickMessagesAddFiltersBtn() .waitUntilAddFiltersMdlVisible() - .setFilterCodeFieldAddFilterMdl(randomAlphanumeric(4)) + .setFilterCodeFieldAddFilterMdl(randomAlphabetic(4)) .selectSaveThisFilterCheckboxMdl(true) .setDisplayNameFldAddFilterMdl(displayName); assertThat(topicDetails.isAddFilterBtnAddFilterMdlEnabled()).as("isAddFilterBtnAddFilterMdlEnabled()") @@ -400,12 +399,41 @@ void checkFilterSavingWithinSavedFilters() { .as("isFilterVisibleAtSavedFiltersMdl()").isTrue(); } + @DisplayName("Checking applying saved filter within Topic/Messages") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(14) + @Test + @Order(12) + void checkingApplyingSavedFilterWithinTopicMessages() { + String displayName = randomAlphabetic(5); + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .openTopic("my_ksql_1ksql_processing_log"); + topicDetails + .waitUntilScreenReady() + .openDetailsTab(MESSAGES) + .clickMessagesAddFiltersBtn() + .waitUntilAddFiltersMdlVisible() + .setFilterCodeFieldAddFilterMdl(randomAlphabetic(4)) + .selectSaveThisFilterCheckboxMdl(true) + .setDisplayNameFldAddFilterMdl(displayName) + .clickAddFilterBtnAndCloseMdl(false) + .openSavedFiltersListMdl() + .selectFilterAtSavedFiltersMdl(displayName) + .clickSelectFilterBtnAtSavedFiltersMdl(); + assertThat(topicDetails.isActiveFilterVisible(displayName)) + .as("isActiveFilterVisible()").isTrue(); + } + @DisplayName("Checking 'Show Internal Topics' toggle functionality within 'All Topics' page") @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) @AutomationStatus(status = Status.AUTOMATED) @CaseId(11) @Test - @Order(12) + @Order(13) void checkShowInternalTopicsButtonFunctionality(){ naviSideBar .openSideMenu(TOPICS);
train
val
2022-12-06T11:13:36
"2022-12-06T10:23:18Z"
anezboretskiy
train
provectus/kafka-ui/3024_3030
provectus/kafka-ui
provectus/kafka-ui/3024
provectus/kafka-ui/3030
[ "connected" ]
43ccca43c2c1aa2cbb6ed9fc97ef99e17daa63bb
1108c760e5f0b23908f3818500b78fe57d44ce71
[ "Hello there Fran-Rg! πŸ‘‹\n\nThank you and congratulations πŸŽ‰ for opening your very first issue in this project! πŸ’–\n\nIn case you want to claim this issue, please comment down below! We will try to get back to you as soon as we can. πŸ‘€", "Other addtional note:\r\nThe service used to work but it's now crashed. There's another service stopping and starting the deployment on a schedule.", "@Fran-Rg hey, thanks for reaching out!\r\nWe've fixed the issue, will merge it today." ]
[]
"2022-12-06T13:06:54Z"
[ "type/bug", "scope/backend", "status/accepted", "status/confirmed" ]
Crashes on IAM auth cluster
**Describe the bug** It will be hard to describe if it's only 1 or many bugs but the cluster ends up with a yellow dot and trying to check the topics crashes the page. There are multiple stack traces in the logs: ``` 2022-12-05 14:26:10,459 WARN [parallel-1] o.a.k.c.u.AppInfoParser: Error registering AppInfo mbean javax.management.InstanceAlreadyExistsException: kafka.admin.client:type=app-info,id=kafka-ui-app at java.management/com.sun.jmx.mbeanserver.Repository.addMBean(Repository.java:436) at java.management/com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerWithRepository(DefaultMBeanServerInterceptor.java:1865) at java.management/com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerDynamicMBean(DefaultMBeanServerInterceptor.java:960) ``` and ``` 2022-12-05 14:26:10,911 INFO [kafka-admin-client-thread | kafka-ui-app] o.a.k.c.NetworkClient: [AdminClient clientId=kafka-ui-app] Node -1 disconnected. 2022-12-05 14:26:10,924 ERROR [scheduling-1] o.s.s.s.TaskUtils$LoggingErrorHandler: Unexpected error occurred in scheduled task java.lang.NullPointerException: null at java.base/java.util.concurrent.ConcurrentHashMap.putVal(ConcurrentHashMap.java:1011) at java.base/java.util.concurrent.ConcurrentHashMap.put(ConcurrentHashMap.java:1006) at com.provectus.kafka.ui.service.StatisticsCache.replace(StatisticsCache.java:26) ``` This is the cluster config: ``` 2022-12-05 14:27:40,452 INFO [parallel-1] o.a.k.c.a.AdminClientConfig: AdminClientConfig values: bootstrap.servers = [<REMOVED>:9098] client.dns.lookup = use_all_dns_ips client.id = kafka-ui-app connections.max.idle.ms = 300000 default.api.timeout.ms = 60000 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retries = 2147483647 retry.backoff.ms = 100 sasl.client.callback.handler.class = class software.amazon.msk.auth.iam.IAMClientCallbackHandler sasl.jaas.config = [hidden] sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = AWS_MSK_IAM sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = SASL_SSL security.providers = null send.buffer.bytes = 131072 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS ``` **Set up** The kafka ui is deployed using helm chart: kafka-ui-0.4.6 This is the values: ``` envs: config: KAFKA_CLUSTERS_0_DISABLELOGDIRSCOLLECTION: "true" KAFKA_CLUSTERS_0_NAME: dev-events KAFKA_CLUSTERS_0_PROPERTIES_SASL_CLIENT_CALLBACK_HANDLER_CLASS: software.amazon.msk.auth.iam.IAMClientCallbackHandler KAFKA_CLUSTERS_0_PROPERTIES_SASL_JAAS_CONFIG: software.amazon.msk.auth.iam.IAMLoginModule required; KAFKA_CLUSTERS_0_PROPERTIES_SASL_MECHANISM: AWS_MSK_IAM KAFKA_CLUSTERS_0_PROPERTIES_SECURITY_PROTOCOL: SASL_SSL KAFKA_CLUSTERS_0_READONLY: "false" SERVER_SERVLET_CONTEXT_PATH: /kafka-ui secret: KAFKA_CLUSTERS_0_BOOTSTRAPSERVERS: <REMOVED>:9098 image: tag: master resources: limits: cpu: "0.25" memory: 1Gi requests: cpu: "0.25" memory: 1Gi serviceAccount: create: false name: kafka-ui ``` **Steps to Reproduce** Steps to reproduce the behavior: 1. Create an aws serverless cluster 2. Deploy the helm with a service account linked to an IAM role granting access to the cluster 3. Check kafka-ui **Expected behavior** UI cluster with a green dot. No stacktraces **Additional context** I used to have a provisioned cluster that doesn't implement any auth and it had never had any issue. The kafka-ui is on: https://github.com/provectus/kafka-ui/commit/43ccca4 (using docker.io/provectuslabs/kafka-ui:master)
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/AdminClientServiceImpl.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ConsumerGroupService.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java" ]
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/AdminClientServiceImpl.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ConsumerGroupService.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java" ]
[ "kafka-ui-api/src/test/java/com/provectus/kafka/ui/service/ReactiveAdminClientTest.java" ]
diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/AdminClientServiceImpl.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/AdminClientServiceImpl.java index 9291198df73..3589a07a47d 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/AdminClientServiceImpl.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/AdminClientServiceImpl.java @@ -37,7 +37,7 @@ private Mono<ReactiveAdminClient> createAdminClient(KafkaCluster cluster) { properties .put(AdminClientConfig.BOOTSTRAP_SERVERS_CONFIG, cluster.getBootstrapServers()); properties.put(AdminClientConfig.REQUEST_TIMEOUT_MS_CONFIG, clientTimeout); - properties.putIfAbsent(AdminClientConfig.CLIENT_ID_CONFIG, "kafka-ui-app"); + properties.putIfAbsent(AdminClientConfig.CLIENT_ID_CONFIG, "kafka-ui-admin-client-" + System.currentTimeMillis()); return AdminClient.create(properties); }) .flatMap(ReactiveAdminClient::create) diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ConsumerGroupService.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ConsumerGroupService.java index 2de70c88ca2..55289a87edd 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ConsumerGroupService.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ConsumerGroupService.java @@ -11,7 +11,6 @@ import java.util.List; import java.util.Map; import java.util.Properties; -import java.util.UUID; import java.util.function.ToIntFunction; import java.util.stream.Collectors; import javax.annotation.Nullable; @@ -197,7 +196,7 @@ public KafkaConsumer<Bytes, Bytes> createConsumer(KafkaCluster cluster, Map<String, Object> properties) { Properties props = new Properties(); props.putAll(cluster.getProperties()); - props.put(ConsumerConfig.CLIENT_ID_CONFIG, "kafka-ui-" + UUID.randomUUID()); + props.put(ConsumerConfig.CLIENT_ID_CONFIG, "kafka-ui-consumer-" + System.currentTimeMillis()); props.put(ConsumerConfig.BOOTSTRAP_SERVERS_CONFIG, cluster.getBootstrapServers()); props.put(ConsumerConfig.KEY_DESERIALIZER_CLASS_CONFIG, BytesDeserializer.class); props.put(ConsumerConfig.VALUE_DESERIALIZER_CLASS_CONFIG, BytesDeserializer.class); diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java index c655cc2ad9b..2e3605145aa 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java @@ -25,9 +25,7 @@ import java.util.Optional; import java.util.Set; import java.util.concurrent.CompletionException; -import java.util.concurrent.ConcurrentHashMap; import java.util.concurrent.ExecutionException; -import java.util.concurrent.atomic.AtomicInteger; import java.util.function.BiFunction; import java.util.function.Function; import java.util.function.Predicate; @@ -260,37 +258,32 @@ public Mono<TopicDescription> describeTopic(String topic) { * such topics in resulting map. * <p/> * This method converts input map into Mono[Map] ignoring keys for which KafkaFutures - * finished with <code>clazz</code> exception. + * finished with <code>clazz</code> exception and empty Monos. */ static <K, V> Mono<Map<K, V>> toMonoWithExceptionFilter(Map<K, KafkaFuture<V>> values, - Class<? extends KafkaException> clazz) { + Class<? extends KafkaException> clazz) { if (values.isEmpty()) { return Mono.just(Map.of()); } - List<Mono<Tuple2<K, V>>> monos = values.entrySet().stream() - .map(e -> toMono(e.getValue()).map(r -> Tuples.of(e.getKey(), r))) - .collect(toList()); - - return Mono.create(sink -> { - var finishedCnt = new AtomicInteger(); - var results = new ConcurrentHashMap<K, V>(); - monos.forEach(mono -> mono.subscribe( - r -> { - results.put(r.getT1(), r.getT2()); - if (finishedCnt.incrementAndGet() == monos.size()) { - sink.success(results); - } - }, - th -> { - if (!th.getClass().isAssignableFrom(clazz)) { - sink.error(th); - } else if (finishedCnt.incrementAndGet() == monos.size()) { - sink.success(results); - } - } - )); - }); + List<Mono<Tuple2<K, Optional<V>>>> monos = values.entrySet().stream() + .map(e -> + toMono(e.getValue()) + .map(r -> Tuples.of(e.getKey(), Optional.of(r))) + .defaultIfEmpty(Tuples.of(e.getKey(), Optional.empty())) //tracking empty Monos + .onErrorResume( + // tracking Monos with suppressible error + th -> th.getClass().isAssignableFrom(clazz), + th -> Mono.just(Tuples.of(e.getKey(), Optional.empty())))) + .toList(); + + return Mono.zip( + monos, + resultsArr -> Stream.of(resultsArr) + .map(obj -> (Tuple2<K, Optional<V>>) obj) + .filter(t -> t.getT2().isPresent()) //skipping empty & suppressible-errors + .collect(Collectors.toMap(Tuple2::getT1, t -> t.getT2().get())) + ); } public Mono<Map<Integer, Map<String, DescribeLogDirsResponse.LogDirInfo>>> describeLogDirs() { @@ -305,6 +298,10 @@ public Mono<Map<Integer, Map<String, DescribeLogDirsResponse.LogDirInfo>>> descr } public Mono<ClusterDescription> describeCluster() { + return describeClusterImpl(client); + } + + private static Mono<ClusterDescription> describeClusterImpl(AdminClient client) { var r = client.describeCluster(); var all = KafkaFuture.allOf(r.nodes(), r.clusterId(), r.controller(), r.authorizedOperations()); return Mono.create(sink -> all.whenComplete((res, ex) -> { @@ -328,15 +325,20 @@ public Mono<ClusterDescription> describeCluster() { } private static Mono<String> getClusterVersion(AdminClient client) { - return toMono(client.describeCluster().controller()) - .flatMap(controller -> loadBrokersConfig(client, List.of(controller.id()))) - .map(configs -> configs.values().stream() + return describeClusterImpl(client) + // choosing node from which we will get configs (starting with controller) + .flatMap(descr -> descr.controller != null + ? Mono.just(descr.controller) + : Mono.justOrEmpty(descr.nodes.stream().findFirst()) + ) + .flatMap(node -> loadBrokersConfig(client, List.of(node.id()))) + .flatMap(configs -> configs.values().stream() .flatMap(Collection::stream) .filter(entry -> entry.name().contains("inter.broker.protocol.version")) .findFirst() - .map(ConfigEntry::value) - .orElse("1.0-UNKNOWN") - ); + .map(configEntry -> Mono.just(configEntry.value())) + .orElse(Mono.empty())) + .switchIfEmpty(Mono.just("1.0-UNKNOWN")); } public Mono<Void> deleteConsumerGroups(Collection<String> groupIds) { @@ -419,6 +421,7 @@ public Mono<Void> alterConsumerGroupOffsets(String groupId, Map<TopicPartition, /** * List offset for the topic's partitions and OffsetSpec. + * * @param failOnUnknownLeader true - throw exception in case of no-leader partitions, * false - skip partitions with no leader */ @@ -432,6 +435,7 @@ public Mono<Map<TopicPartition, Long>> listTopicOffsets(String topic, /** * List offset for the specified partitions and OffsetSpec. + * * @param failOnUnknownLeader true - throw exception in case of no-leader partitions, * false - skip partitions with no leader */
diff --git a/kafka-ui-api/src/test/java/com/provectus/kafka/ui/service/ReactiveAdminClientTest.java b/kafka-ui-api/src/test/java/com/provectus/kafka/ui/service/ReactiveAdminClientTest.java index 26df31426da..99cfedad4cf 100644 --- a/kafka-ui-api/src/test/java/com/provectus/kafka/ui/service/ReactiveAdminClientTest.java +++ b/kafka-ui-api/src/test/java/com/provectus/kafka/ui/service/ReactiveAdminClientTest.java @@ -104,7 +104,14 @@ void testToMonoWithExceptionFilter() { var okFuture = new KafkaFutureImpl<String>(); okFuture.complete("done"); - Map<String, KafkaFuture<String>> arg = Map.of("failure", failedFuture, "ok", okFuture); + var emptyFuture = new KafkaFutureImpl<String>(); + emptyFuture.complete(null); + + Map<String, KafkaFuture<String>> arg = Map.of( + "failure", failedFuture, + "ok", okFuture, + "empty", emptyFuture + ); StepVerifier.create(toMonoWithExceptionFilter(arg, UnknownTopicOrPartitionException.class)) .assertNext(result -> assertThat(result).hasSize(1).containsEntry("ok", "done")) .verifyComplete();
train
val
2022-12-05T14:02:45
"2022-12-05T14:37:56Z"
Fran-Rg
train
provectus/kafka-ui/3015_3045
provectus/kafka-ui
provectus/kafka-ui/3015
provectus/kafka-ui/3045
[ "connected" ]
53fb641220aaa86c6f3bf0a4521748e039ca7a4e
c13440b65d602399d27d7110580fe18856ad0603
[]
[ "navigateToConnectorsAndOpenConnectorsDetails -> navigateToConnectorsAndOpenDetails", "here u can also reuse created navigateToConnectors()", "navigateToSchema -> navigateToSchemaRegistry", "navigateToSchemaAndOpenSchemaDetails -> navigateToSchemaRegistryAndOpenDetails", "here u can also reuse created navigateToSchema()", "seems here u forgot to use navigateToTopics()", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "Fixed", "navigateToSchemaRegistryAndOpenSchemaDetails -> navigateToSchemaRegistryAndOpenDetails", "use here also verifyEnabledElements()", "use at 89 line also verifyVisibleElements()", "suggest to update verifyElementsCondition() and reuse here", "let's merge this two methods by using Condition as an argument", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "getAllMenuButtons() list should be visible and enabled at same time\njust add it in 16 and 17 line", "Fixed." ]
"2022-12-08T12:26:23Z"
[ "scope/QA", "scope/AQA" ]
[e2e] Create method with common steps inside test classes
Almost every test in concrete class starts from the same steps. In case they aren't main goal of those tests let's create private void method with Step annotation containing common steps and reuse it.
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersList.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersList.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/BaseTest.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/SmokeTests.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/connectors/ConnectorsTests.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/schemas/SchemasTests.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicMessagesTests.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersList.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersList.java index 3df1a09304b..98ed497e422 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersList.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersList.java @@ -30,7 +30,7 @@ public BrokersList openBroker(String brokerId) { } private List<SelenideElement> getUptimeSummaryCells() { - return Stream.of("Broker Count", "Active Controllers", "Version") + return Stream.of("Broker Count", "Active Controller", "Version") .map(name -> $x(String.format(summaryCellLocator, name))) .collect(Collectors.toList()); } @@ -87,7 +87,7 @@ public BrokerGridItem(SelenideElement element) { } private SelenideElement getIdElm() { - return element.$x("./td[1]/a"); + return element.$x("./td[1]/div/a"); } @Step
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/BaseTest.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/BaseTest.java index 11cf07c622e..d3a279cc713 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/BaseTest.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/BaseTest.java @@ -1,13 +1,17 @@ package com.provectus.kafka.ui.base; +import com.codeborne.selenide.Condition; import com.codeborne.selenide.Selenide; +import com.codeborne.selenide.SelenideElement; import com.codeborne.selenide.WebDriverRunner; import com.provectus.kafka.ui.utilities.qaseIoUtils.DisplayNameGenerator; import com.provectus.kafka.ui.utilities.qaseIoUtils.TestCaseGenerator; import io.github.cdimascio.dotenv.Dotenv; import io.qameta.allure.Allure; +import io.qase.api.annotation.Step; import lombok.extern.slf4j.Slf4j; import org.apache.commons.io.FileUtils; +import org.assertj.core.api.SoftAssertions; import org.junit.jupiter.api.*; import org.openqa.selenium.Dimension; import org.openqa.selenium.OutputType; @@ -22,8 +26,10 @@ import java.io.ByteArrayInputStream; import java.io.File; import java.io.IOException; +import java.util.List; import static com.provectus.kafka.ui.base.Setup.*; +import static com.provectus.kafka.ui.pages.NaviSideBar.SideMenuOption.TOPICS; import static com.provectus.kafka.ui.settings.Source.BASE_WEB_URL; @Slf4j @@ -110,4 +116,31 @@ public void afterMethod() { ((TakesScreenshot) webDriverContainer.getWebDriver()).getScreenshotAs(OutputType.BYTES))); browserClear(); } + + @Step + protected void navigateToTopics(){ + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady(); + } + + @Step + protected void navigateToTopicsAndOpenDetails(String topicName){ + naviSideBar + .openSideMenu(TOPICS); + topicsList + .waitUntilScreenReady() + .openTopic(topicName); + topicDetails + .waitUntilScreenReady(); + } + + @Step + protected void verifyElementsCondition(List<SelenideElement> elementList, Condition expectedCondition) { + SoftAssertions softly = new SoftAssertions(); + elementList.forEach(element -> softly.assertThat(element.is(expectedCondition)) + .as(element.getSearchCriteria() + " is " + expectedCondition).isTrue()); + softly.assertAll(); + } } diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/SmokeTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/SmokeTests.java index ed7d901fe4b..ac5f5c57557 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/SmokeTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/SmokeTests.java @@ -5,7 +5,8 @@ import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.AutomationStatus; import com.provectus.kafka.ui.utilities.qaseIoUtils.enums.Status; import io.qase.api.annotation.CaseId; -import org.assertj.core.api.SoftAssertions; +import java.util.stream.Collectors; +import java.util.stream.Stream; import org.junit.jupiter.api.Test; public class SmokeTests extends BaseTest { @@ -14,19 +15,9 @@ public class SmokeTests extends BaseTest { @AutomationStatus(status = Status.AUTOMATED) @CaseId(198) public void checkBasePageElements(){ - SoftAssertions softly = new SoftAssertions(); - topPanel.getAllVisibleElements() - .forEach(element -> - softly.assertThat(element.is(Condition.visible)) - .as(element.getSearchCriteria() + " isVisible()").isTrue()); - topPanel.getAllEnabledElements() - .forEach(element -> - softly.assertThat(element.is(Condition.enabled)) - .as(element.getSearchCriteria() + " isEnabled()").isTrue()); - naviSideBar.getAllMenuButtons() - .forEach(element -> - softly.assertThat(element.is(Condition.enabled) && element.is(Condition.visible)) - .as(element.getSearchCriteria() + " isEnabled()").isTrue()); - softly.assertAll(); + verifyElementsCondition(Stream.concat(topPanel.getAllVisibleElements().stream(), naviSideBar.getAllMenuButtons().stream()) + .collect(Collectors.toList()),Condition.visible); + verifyElementsCondition(Stream.concat(topPanel.getAllEnabledElements().stream(), naviSideBar.getAllMenuButtons().stream()) + .collect(Collectors.toList()),Condition.enabled); } } diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java index dcff92bc28c..0b54a4015ce 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java @@ -8,8 +8,8 @@ import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.AutomationStatus; import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.Suite; import com.provectus.kafka.ui.utilities.qaseIoUtils.enums.Status; +import io.qameta.allure.Step; import io.qase.api.annotation.CaseId; -import org.assertj.core.api.SoftAssertions; import org.junit.jupiter.api.DisplayName; import org.junit.jupiter.api.Test; @@ -23,17 +23,10 @@ public class BrokersTests extends BaseTest { @CaseId(1) @Test public void checkBrokersOverview(){ - naviSideBar - .openSideMenu(BROKERS); - brokersList - .waitUntilScreenReady(); - SoftAssertions softly = new SoftAssertions(); - softly.assertThat(brokersList.getAllBrokers()).as("getAllBrokers()").size().isGreaterThan(0); - brokersList.getAllVisibleElements().forEach(element -> softly.assertThat(element.is(Condition.visible)) - .as(element.getSearchCriteria() + " isVisible()").isTrue()); - brokersList.getAllEnabledElements().forEach(element -> softly.assertThat(element.is(Condition.enabled)) - .as(element.getSearchCriteria() + " isEnabled()").isTrue()); - softly.assertAll(); + navigateToBrokers(); + assertThat(brokersList.getAllBrokers()).as("getAllBrokers()").size().isGreaterThan(0); + verifyElementsCondition(brokersList.getAllVisibleElements(), Condition.visible); + verifyElementsCondition(brokersList.getAllEnabledElements(), Condition.enabled); } @DisplayName("Checking the existing Broker's profile in a cluster") @@ -42,20 +35,21 @@ public void checkBrokersOverview(){ @CaseId(85) @Test public void checkExistingBrokersInCluster(){ - naviSideBar - .openSideMenu(BROKERS); - brokersList - .waitUntilScreenReady(); + navigateToBrokers(); assertThat(brokersList.getAllBrokers()).as("getAllBrokers()").size().isGreaterThan(0); brokersList .openBroker("1"); brokersDetails .waitUntilScreenReady(); - SoftAssertions softly = new SoftAssertions(); - brokersDetails.getAllVisibleElements().forEach(element -> softly.assertThat(element.is(Condition.visible)) - .as(element.getSearchCriteria() + " isVisible()").isTrue()); - brokersDetails.getAllEnabledElements().forEach(element -> softly.assertThat(element.is(Condition.enabled)) - .as(element.getSearchCriteria() + " isEnabled()").isTrue()); - softly.assertAll(); + verifyElementsCondition(brokersDetails.getAllVisibleElements(), Condition.visible); + verifyElementsCondition(brokersDetails.getAllVisibleElements(), Condition.enabled); + } + + @Step + private void navigateToBrokers(){ + naviSideBar + .openSideMenu(BROKERS); + brokersList + .waitUntilScreenReady(); } } diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/connectors/ConnectorsTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/connectors/ConnectorsTests.java index 5e8ed9157ab..980c2d9d48b 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/connectors/ConnectorsTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/connectors/ConnectorsTests.java @@ -12,6 +12,7 @@ import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.AutomationStatus; import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.Suite; import com.provectus.kafka.ui.utilities.qaseIoUtils.enums.Status; +import io.qameta.allure.Step; import io.qase.api.annotation.CaseId; import java.util.ArrayList; import java.util.List; @@ -68,10 +69,8 @@ public void createConnector() { Connector connectorForCreate = new Connector() .setName("sink_postgres_activities_e2e_checks-" + randomAlphabetic(5)) .setConfig(getResourceAsString("config_for_create_connector.json")); - naviSideBar - .openSideMenu(KAFKA_CONNECT); - kafkaConnectList - .waitUntilScreenReady() + navigateToConnectors(); + kafkaConnectList .clickCreateConnectorBtn(); connectorCreateForm .waitUntilScreenReady() @@ -79,18 +78,9 @@ public void createConnector() { .clickSubmitButton(); connectorDetails .waitUntilScreenReady(); - naviSideBar - .openSideMenu(KAFKA_CONNECT); - kafkaConnectList - .waitUntilScreenReady() - .openConnector(connectorForCreate.getName()); - connectorDetails - .waitUntilScreenReady(); + navigateToConnectorsAndOpenDetails(connectorForCreate.getName()); Assertions.assertTrue(connectorDetails.isConnectorHeaderVisible(connectorForCreate.getName()),"isConnectorTitleVisible()"); - naviSideBar - .openSideMenu(KAFKA_CONNECT); - kafkaConnectList - .waitUntilScreenReady(); + navigateToConnectors(); Assertions.assertTrue(kafkaConnectList.isConnectorVisible(CONNECTOR_FOR_DELETE.getName()), "isConnectorVisible()"); CONNECTOR_LIST.add(connectorForCreate); } @@ -101,21 +91,13 @@ public void createConnector() { @CaseId(196) @Test public void updateConnector() { - naviSideBar - .openSideMenu(KAFKA_CONNECT); - kafkaConnectList - .waitUntilScreenReady() - .openConnector(CONNECTOR_FOR_UPDATE.getName()); - connectorDetails - .waitUntilScreenReady() + navigateToConnectorsAndOpenDetails(CONNECTOR_FOR_UPDATE.getName()); + connectorDetails .openConfigTab() .setConfig(CONNECTOR_FOR_UPDATE.getConfig()) .clickSubmitButton(); Assertions.assertTrue(connectorDetails.isAlertWithMessageVisible(SUCCESS,"Config successfully updated."),"isAlertWithMessageVisible()"); - naviSideBar - .openSideMenu(KAFKA_CONNECT); - kafkaConnectList - .waitUntilScreenReady(); + navigateToConnectors(); Assertions.assertTrue(kafkaConnectList.isConnectorVisible(CONNECTOR_FOR_UPDATE.getName()), "isConnectorVisible()"); } @@ -125,20 +107,12 @@ public void updateConnector() { @CaseId(195) @Test public void deleteConnector() { - naviSideBar - .openSideMenu(KAFKA_CONNECT); - kafkaConnectList - .waitUntilScreenReady() - .openConnector(CONNECTOR_FOR_DELETE.getName()); - connectorDetails - .waitUntilScreenReady() + navigateToConnectorsAndOpenDetails(CONNECTOR_FOR_DELETE.getName()); + connectorDetails .openDotMenu() .clickDeleteBtn() .clickConfirmBtn(); - naviSideBar - .openSideMenu(KAFKA_CONNECT); - kafkaConnectList - .waitUntilScreenReady(); + navigateToConnectors(); Assertions.assertFalse(kafkaConnectList.isConnectorVisible(CONNECTOR_FOR_DELETE.getName()), "isConnectorVisible()"); CONNECTOR_LIST.remove(CONNECTOR_FOR_DELETE); } @@ -149,4 +123,21 @@ public void afterAll() { apiHelper.deleteConnector(CLUSTER_NAME, CONNECT_NAME, connector.getName())); TOPIC_LIST.forEach(topic -> apiHelper.deleteTopic(CLUSTER_NAME, topic.getName())); } + + @Step + private void navigateToConnectors(){ + naviSideBar + .openSideMenu(KAFKA_CONNECT); + kafkaConnectList + .waitUntilScreenReady(); + } + + @Step + private void navigateToConnectorsAndOpenDetails(String connectorName){ + navigateToConnectors(); + kafkaConnectList + .openConnector(connectorName); + connectorDetails + .waitUntilScreenReady(); + } } diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/schemas/SchemasTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/schemas/SchemasTests.java index f112c7b5881..92daffd2495 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/schemas/SchemasTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/schemas/SchemasTests.java @@ -11,6 +11,7 @@ import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.AutomationStatus; import com.provectus.kafka.ui.utilities.qaseIoUtils.annotations.Suite; import com.provectus.kafka.ui.utilities.qaseIoUtils.enums.Status; +import io.qameta.allure.Step; import io.qase.api.annotation.CaseId; import java.util.ArrayList; import java.util.List; @@ -51,10 +52,8 @@ public void beforeAll() { @Order(1) void createSchemaAvro() { Schema schemaAvro = Schema.createSchemaAvro(); - naviSideBar - .openSideMenu(SCHEMA_REGISTRY); + navigateToSchemaRegistry(); schemaRegistryList - .waitUntilScreenReady() .clickCreateSchema(); schemaCreateForm .setSubjectName(schemaAvro.getName()) @@ -68,10 +67,7 @@ void createSchemaAvro() { softly.assertThat(schemaDetails.getSchemaType()).as("getSchemaType()").isEqualTo(schemaAvro.getType().getValue()); softly.assertThat(schemaDetails.getCompatibility()).as("getCompatibility()").isEqualTo(CompatibilityLevel.CompatibilityEnum.BACKWARD.getValue()); softly.assertAll(); - naviSideBar - .openSideMenu(SCHEMA_REGISTRY); - schemaRegistryList - .waitUntilScreenReady(); + navigateToSchemaRegistry(); Assertions.assertTrue(schemaRegistryList.isSchemaVisible(AVRO_API.getName()),"isSchemaVisible()"); SCHEMA_LIST.add(schemaAvro); } @@ -84,20 +80,13 @@ void createSchemaAvro() { @Order(2) void updateSchemaAvro() { AVRO_API.setValuePath(System.getProperty("user.dir") + "/src/main/resources/testData/schema_avro_for_update.json"); - naviSideBar - .openSideMenu(SCHEMA_REGISTRY); - schemaRegistryList - .waitUntilScreenReady() - .openSchema(AVRO_API.getName()); + navigateToSchemaRegistryAndOpenDetails(AVRO_API.getName()); schemaDetails - .waitUntilScreenReady() .openEditSchema(); schemaCreateForm .waitUntilScreenReady(); - SoftAssertions softly = new SoftAssertions(); - schemaCreateForm.getAllDetailsPageElements() - .forEach(element -> softly.assertThat(element.is(Condition.visible)) - .as(element.getSearchCriteria() + " isVisible").isTrue()); + verifyElementsCondition(schemaCreateForm.getAllDetailsPageElements(), Condition.visible); + SoftAssertions softly = new SoftAssertions(); softly.assertThat(schemaCreateForm.isSubmitBtnEnabled()).as("isSubmitBtnEnabled()").isFalse(); softly.assertThat(schemaCreateForm.isSchemaDropDownEnabled()).as("isSchemaDropDownEnabled()").isFalse(); softly.assertAll(); @@ -117,11 +106,7 @@ void updateSchemaAvro() { @Test @Order(3) void compareVersionsOperation() { - naviSideBar - .openSideMenu(SCHEMA_REGISTRY); - schemaRegistryList - .waitUntilScreenReady() - .openSchema(AVRO_API.getName()); + navigateToSchemaRegistryAndOpenDetails(AVRO_API.getName()); int latestVersion = schemaDetails .waitUntilScreenReady() .getLatestVersion(); @@ -144,13 +129,8 @@ void compareVersionsOperation() { @Test @Order(4) void deleteSchemaAvro() { - naviSideBar - .openSideMenu(SCHEMA_REGISTRY); - schemaRegistryList - .waitUntilScreenReady() - .openSchema(AVRO_API.getName()); + navigateToSchemaRegistryAndOpenDetails(AVRO_API.getName()); schemaDetails - .waitUntilScreenReady() .removeSchema(); schemaRegistryList .waitUntilScreenReady(); @@ -166,10 +146,8 @@ void deleteSchemaAvro() { @Order(5) void createSchemaJson() { Schema schemaJson = Schema.createSchemaJson(); - naviSideBar - .openSideMenu(SCHEMA_REGISTRY); + navigateToSchemaRegistry(); schemaRegistryList - .waitUntilScreenReady() .clickCreateSchema(); schemaCreateForm .setSubjectName(schemaJson.getName()) @@ -183,10 +161,7 @@ void createSchemaJson() { softly.assertThat(schemaDetails.getSchemaType()).as("getSchemaType()").isEqualTo(schemaJson.getType().getValue()); softly.assertThat(schemaDetails.getCompatibility()).as("getCompatibility()").isEqualTo(CompatibilityLevel.CompatibilityEnum.BACKWARD.getValue()); softly.assertAll(); - naviSideBar - .openSideMenu(SCHEMA_REGISTRY); - schemaRegistryList - .waitUntilScreenReady(); + navigateToSchemaRegistry(); Assertions.assertTrue(schemaRegistryList.isSchemaVisible(JSON_API.getName()),"isSchemaVisible()"); SCHEMA_LIST.add(schemaJson); } @@ -198,13 +173,8 @@ void createSchemaJson() { @Test @Order(6) void deleteSchemaJson() { - naviSideBar - .openSideMenu(SCHEMA_REGISTRY); - schemaRegistryList - .waitUntilScreenReady() - .openSchema(JSON_API.getName()); - schemaDetails - .waitUntilScreenReady() + navigateToSchemaRegistryAndOpenDetails(JSON_API.getName()); + schemaDetails .removeSchema(); schemaRegistryList .waitUntilScreenReady(); @@ -220,10 +190,8 @@ void deleteSchemaJson() { @Order(7) void createSchemaProtobuf() { Schema schemaProtobuf = Schema.createSchemaProtobuf(); - naviSideBar - .openSideMenu(SCHEMA_REGISTRY); + navigateToSchemaRegistry(); schemaRegistryList - .waitUntilScreenReady() .clickCreateSchema(); schemaCreateForm .setSubjectName(schemaProtobuf.getName()) @@ -237,10 +205,7 @@ void createSchemaProtobuf() { softly.assertThat(schemaDetails.getSchemaType()).as("getSchemaType()").isEqualTo(schemaProtobuf.getType().getValue()); softly.assertThat(schemaDetails.getCompatibility()).as("getCompatibility()").isEqualTo(CompatibilityLevel.CompatibilityEnum.BACKWARD.getValue()); softly.assertAll(); - naviSideBar - .openSideMenu(SCHEMA_REGISTRY); - schemaRegistryList - .waitUntilScreenReady(); + navigateToSchemaRegistry(); Assertions.assertTrue(schemaRegistryList.isSchemaVisible(PROTOBUF_API.getName()),"isSchemaVisible()"); SCHEMA_LIST.add(schemaProtobuf); } @@ -252,13 +217,8 @@ void createSchemaProtobuf() { @Test @Order(8) void deleteSchemaProtobuf() { - naviSideBar - .openSideMenu(SCHEMA_REGISTRY); - schemaRegistryList - .waitUntilScreenReady() - .openSchema(PROTOBUF_API.getName()); - schemaDetails - .waitUntilScreenReady() + navigateToSchemaRegistryAndOpenDetails(PROTOBUF_API.getName()); + schemaDetails .removeSchema(); schemaRegistryList .waitUntilScreenReady(); @@ -270,4 +230,21 @@ void deleteSchemaProtobuf() { public void afterAll() { SCHEMA_LIST.forEach(schema -> apiHelper.deleteSchema(CLUSTER_NAME, schema.getName())); } + + @Step + private void navigateToSchemaRegistry(){ + naviSideBar + .openSideMenu(SCHEMA_REGISTRY); + schemaRegistryList + .waitUntilScreenReady(); + } + + @Step + private void navigateToSchemaRegistryAndOpenDetails(String schemaName){ + navigateToSchemaRegistry(); + schemaRegistryList + .openSchema(schemaName); + schemaDetails + .waitUntilScreenReady(); + } } diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicMessagesTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicMessagesTests.java index f3da20747be..8fbf915d543 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicMessagesTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicMessagesTests.java @@ -47,13 +47,8 @@ public void beforeAll() { @CaseId(222) @Test void produceMessage() { - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady() - .openTopic(TOPIC_FOR_MESSAGES.getName()); + navigateToTopicsAndOpenDetails(TOPIC_FOR_MESSAGES.getName()); topicDetails - .waitUntilScreenReady() .openDetailsTab(TopicDetails.TopicMenu.MESSAGES) .clickProduceMessageBtn(); produceMessagePanel @@ -77,13 +72,8 @@ void produceMessage() { @CaseId(19) @Test void clearMessage() { - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady() - .openTopic(TOPIC_FOR_MESSAGES.getName()); + navigateToTopicsAndOpenDetails(TOPIC_FOR_MESSAGES.getName()); topicDetails - .waitUntilScreenReady() .openDetailsTab(TopicDetails.TopicMenu.OVERVIEW) .clickProduceMessageBtn(); int messageAmount = topicDetails.getMessageCountAmount(); @@ -111,13 +101,8 @@ void clearMessage() { @Test void copyMessageFromTopicProfile() { String topicName = "_schemas"; - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady() - .openTopic(topicName); + navigateToTopicsAndOpenDetails(topicName); topicDetails - .waitUntilScreenReady() .openDetailsTab(TopicDetails.TopicMenu.MESSAGES) .getRandomMessage() .openDotMenu() diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java index fbee8c82610..e75c39f7e97 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java @@ -1,6 +1,5 @@ package com.provectus.kafka.ui.suite.topics; -import static com.provectus.kafka.ui.pages.NaviSideBar.SideMenuOption.TOPICS; import static com.provectus.kafka.ui.pages.topic.TopicDetails.TopicMenu.MESSAGES; import static com.provectus.kafka.ui.pages.topic.TopicDetails.TopicMenu.SETTINGS; import static com.provectus.kafka.ui.pages.topic.enums.CleanupPolicyValue.COMPACT; @@ -72,10 +71,8 @@ public void beforeAll() { @Test @Order(1) public void createTopic() { - naviSideBar - .openSideMenu(TOPICS); + navigateToTopics(); topicsList - .waitUntilScreenReady() .clickAddTopicBtn(); topicCreateEditForm .waitUntilScreenReady() @@ -83,13 +80,7 @@ public void createTopic() { .setNumberOfPartitions(TOPIC_TO_CREATE.getNumberOfPartitions()) .selectCleanupPolicy(TOPIC_TO_CREATE.getCleanupPolicyValue()) .clickCreateTopicBtn(); - topicDetails - .waitUntilScreenReady(); - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady() - .openTopic(TOPIC_TO_CREATE.getName()); + navigateToTopicsAndOpenDetails(TOPIC_TO_CREATE.getName()); SoftAssertions softly = new SoftAssertions(); softly.assertThat(topicDetails.isTopicHeaderVisible(TOPIC_TO_CREATE.getName())).as("isTopicHeaderVisible()") .isTrue(); @@ -98,10 +89,7 @@ public void createTopic() { softly.assertThat(topicDetails.getPartitions()).as("getPartitions()") .isEqualTo(TOPIC_TO_CREATE.getNumberOfPartitions()); softly.assertAll(); - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady(); + navigateToTopics(); Assertions.assertTrue(topicsList.isTopicVisible(TOPIC_TO_CREATE.getName()), "isTopicVisible"); TOPIC_LIST.add(TOPIC_TO_CREATE); } @@ -113,16 +101,11 @@ public void createTopic() { @Test @Order(2) void checkAvailableOperations() { - naviSideBar - .openSideMenu(TOPICS); + navigateToTopics(); topicsList - .waitUntilScreenReady() .getTopicItem("my_ksql_1ksql_processing_log") .selectItem(true); - topicsList - .getActionButtons() - .forEach(element -> assertThat(element.is(Condition.enabled)) - .as(element.getSearchCriteria() + " isEnabled()").isTrue()); + verifyElementsCondition(topicsList.getActionButtons(),Condition.enabled); topicsList .getTopicItem("_confluent-ksql-my_ksql_1_command_topic") .selectItem(true); @@ -138,13 +121,8 @@ void checkAvailableOperations() { @Test @Order(3) public void updateTopic() { - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady() - .openTopic(TOPIC_FOR_UPDATE.getName()); + navigateToTopicsAndOpenDetails(TOPIC_FOR_UPDATE.getName()); topicDetails - .waitUntilScreenReady() .openDotMenu() .clickEditSettingsMenu(); topicCreateEditForm @@ -157,13 +135,8 @@ public void updateTopic() { .clickCreateTopicBtn(); topicDetails .waitUntilScreenReady(); - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady() - .openTopic(TOPIC_FOR_UPDATE.getName()); + navigateToTopicsAndOpenDetails(TOPIC_FOR_UPDATE.getName()); topicDetails - .waitUntilScreenReady() .openDotMenu() .clickEditSettingsMenu(); SoftAssertions softly = new SoftAssertions(); @@ -185,20 +158,12 @@ public void updateTopic() { @Test @Order(4) public void deleteTopic() { - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady() - .openTopic(TOPIC_FOR_DELETE.getName()); + navigateToTopicsAndOpenDetails(TOPIC_FOR_DELETE.getName()); topicDetails - .waitUntilScreenReady() .openDotMenu() .clickDeleteTopicMenu() .clickConfirmDeleteBtn(); - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady(); + navigateToTopics(); Assertions.assertFalse(topicsList.isTopicVisible(TOPIC_FOR_DELETE.getName()), "isTopicVisible"); TOPIC_LIST.remove(TOPIC_FOR_DELETE); } @@ -212,13 +177,8 @@ public void deleteTopic() { void redirectToConsumerFromTopic() { String topicName = "source-activities"; String consumerGroupId = "connect-sink_postgres_activities"; - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady() - .openTopic(topicName); + navigateToTopicsAndOpenDetails(topicName); topicDetails - .waitUntilScreenReady() .openDetailsTab(TopicDetails.TopicMenu.CONSUMERS) .openConsumerGroup(consumerGroupId); consumersDetails @@ -236,10 +196,8 @@ void redirectToConsumerFromTopic() { @Test @Order(6) void checkTopicCreatePossibility() { - naviSideBar - .openSideMenu(TOPICS); + navigateToTopics(); topicsList - .waitUntilScreenReady() .clickAddTopicBtn(); topicCreateEditForm .waitUntilScreenReady(); @@ -266,14 +224,14 @@ void checkTimeToRetainDataCustomValueWithEditingTopic() { Topic topicToRetainData = new Topic() .setName("topic-to-retain-data-" + randomAlphabetic(5)) .setTimeToRetainData("86400000"); - naviSideBar - .openSideMenu(TOPICS); + navigateToTopics(); topicsList - .waitUntilScreenReady() .clickAddTopicBtn(); topicCreateEditForm .waitUntilScreenReady() - .setTopicName(topicToRetainData.getName()); + .setTopicName(topicToRetainData.getName()) + .setNumberOfPartitions(1) + .setTimeToRetainDataInMs("604800000"); assertThat(topicCreateEditForm.getTimeToRetain()).as("getTimeToRetain()").isEqualTo("604800000"); topicCreateEditForm .setTimeToRetainDataInMs(topicToRetainData.getTimeToRetainData()) @@ -298,8 +256,7 @@ void checkTimeToRetainDataCustomValueWithEditingTopic() { @Test @Order(8) void checkCustomParametersWithinCreateNewTopic() { - naviSideBar - .openSideMenu(TOPICS); + navigateToTopics(); topicsList .waitUntilScreenReady() .clickAddTopicBtn(); @@ -323,18 +280,9 @@ void checkCustomParametersWithinCreateNewTopic() { @Test @Order(9) void checkTopicListElements() { - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady(); - SoftAssertions softly = new SoftAssertions(); - topicsList.getAllVisibleElements().forEach( - element -> softly.assertThat(element.is(Condition.visible)).as(element.getSearchCriteria() + " isVisible()") - .isTrue()); - topicsList.getAllEnabledElements().forEach( - element -> softly.assertThat(element.is(Condition.enabled)).as(element.getSearchCriteria() + " isEnabled()") - .isTrue()); - softly.assertAll(); + navigateToTopics(); + verifyElementsCondition(topicsList.getAllVisibleElements(), Condition.visible); + verifyElementsCondition(topicsList.getAllEnabledElements(), Condition.enabled); } @DisplayName("Filter adding within Topic") @@ -345,29 +293,16 @@ void checkTopicListElements() { @Order(10) void addingNewFilterWithinTopic() { String filterName = randomAlphabetic(5); - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady() - .openTopic("_schemas"); + navigateToTopicsAndOpenDetails("_schemas"); topicDetails - .waitUntilScreenReady() .openDetailsTab(MESSAGES) .clickMessagesAddFiltersBtn() .waitUntilAddFiltersMdlVisible(); - SoftAssertions softly = new SoftAssertions(); - topicDetails.getAllAddFilterModalVisibleElements().forEach(element -> - softly.assertThat(element.is(Condition.visible)) - .as(element.getSearchCriteria() + " isVisible()").isTrue()); - topicDetails.getAllAddFilterModalEnabledElements().forEach(element -> - softly.assertThat(element.is(Condition.enabled)) - .as(element.getSearchCriteria() + " isEnabled()").isTrue()); - topicDetails.getAllAddFilterModalDisabledElements().forEach(element -> - softly.assertThat(element.is(Condition.enabled)) - .as(element.getSearchCriteria() + " isEnabled()").isFalse()); - softly.assertThat(topicDetails.isSaveThisFilterCheckBoxSelected()).as("isSaveThisFilterCheckBoxSelected()") + verifyElementsCondition(topicDetails.getAllAddFilterModalVisibleElements(), Condition.visible); + verifyElementsCondition(topicDetails.getAllAddFilterModalEnabledElements(), Condition.enabled); + verifyElementsCondition(topicDetails.getAllAddFilterModalDisabledElements(), Condition.disabled); + assertThat(topicDetails.isSaveThisFilterCheckBoxSelected()).as("isSaveThisFilterCheckBoxSelected()") .isFalse(); - softly.assertAll(); topicDetails .setFilterCodeFieldAddFilterMdl(filterName); assertThat(topicDetails.isAddFilterBtnAddFilterMdlEnabled()).as("isAddFilterBtnAddFilterMdlEnabled()") @@ -385,13 +320,8 @@ void addingNewFilterWithinTopic() { @Order(11) void checkFilterSavingWithinSavedFilters() { String displayName = randomAlphabetic(5); - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady() - .openTopic("my_ksql_1ksql_processing_log"); + navigateToTopicsAndOpenDetails("my_ksql_1ksql_processing_log"); topicDetails - .waitUntilScreenReady() .openDetailsTab(MESSAGES) .clickMessagesAddFiltersBtn() .waitUntilAddFiltersMdlVisible() @@ -415,13 +345,8 @@ void checkFilterSavingWithinSavedFilters() { @Order(12) void checkingApplyingSavedFilterWithinTopicMessages() { String displayName = randomAlphabetic(5); - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady() - .openTopic("my_ksql_1ksql_processing_log"); + navigateToTopicsAndOpenDetails("my_ksql_1ksql_processing_log"); topicDetails - .waitUntilScreenReady() .openDetailsTab(MESSAGES) .clickMessagesAddFiltersBtn() .waitUntilAddFiltersMdlVisible() @@ -443,10 +368,7 @@ void checkingApplyingSavedFilterWithinTopicMessages() { @Test @Order(13) void checkShowInternalTopicsButtonFunctionality(){ - naviSideBar - .openSideMenu(TOPICS); - topicsList - .waitUntilScreenReady(); + navigateToTopics(); SoftAssertions softly = new SoftAssertions(); softly.assertThat(topicsList.isShowInternalRadioBtnSelected()).as("isInternalRadioBtnSelected()").isTrue(); softly.assertThat(topicsList.getInternalTopics()).as("getInternalTopics()").size().isGreaterThan(0);
train
val
2022-12-12T16:42:38
"2022-12-02T14:44:03Z"
VladSenyuta
train
provectus/kafka-ui/2664_3049
provectus/kafka-ui
provectus/kafka-ui/2664
provectus/kafka-ui/3049
[ "connected" ]
4dc0f6d81e90b384e6eda1f69bde892dfd66a305
ddc76ac3ba1ce6e20357ee132b7e9e1bedba0ef3
[]
[]
"2022-12-09T11:32:27Z"
[ "status/accepted", "scope/infrastructure" ]
Modify a release flow to publish serde-spi jar to maven central
[]
[ ".github/workflows/release-serde-api.yaml" ]
[]
diff --git a/.github/workflows/release-serde-api.yaml b/.github/workflows/release-serde-api.yaml new file mode 100644 index 00000000000..e13fa7a23a0 --- /dev/null +++ b/.github/workflows/release-serde-api.yaml @@ -0,0 +1,33 @@ +name: Release-serde-api +on: workflow_dispatch + +jobs: + release-serde-api: + runs-on: ubuntu-latest + steps: + - uses: actions/checkout@v3 + with: + fetch-depth: 0 + + - run: | + git config user.name github-actions + git config user.email [email protected] + + - name: Set up JDK + uses: actions/setup-java@v3 + with: + java-version: "17" + distribution: "zulu" + cache: "maven" + + - id: install-secret-key + name: Install GPG secret key + run: | + cat <(echo -e "${{ secrets.GPG_PRIVATE_KEY }}") | gpg --batch --import + + - name: Publish to Maven Central + run: | + MVN_VERSION=$(curl -s https://search.maven.org/solrsearch/select?q=g:"com.provectus"+AND+a:"kafka-ui-serde-api" | grep -o '"latestVersion": *"[^"]*"' | grep -o '"[^"]*"$' | sed 's/"//g') + MVN_VERSION=$(echo "$MVN_VERSION" | awk 'BEGIN{FS=OFS="."} {$2+=1} 1') + mvn -B -ntp versions:set -DnewVersion=$MVN_VERSION -pl kafka-ui-serde-api + mvn source:jar javadoc:jar package gpg:sign -Dgpg.passphrase=${{ secrets.GPG_PASSPHRASE }} -Dserver.username=${{ secrets.NEXUS_USERNAME }} -Dserver.password=${{ secrets.NEXUS_PASSWORD }} nexus-staging:deploy -pl kafka-ui-serde-api -s settings.xml
null
train
val
2022-12-09T10:31:51
"2022-09-28T06:16:54Z"
Haarolean
train
provectus/kafka-ui/2624_3057
provectus/kafka-ui
provectus/kafka-ui/2624
provectus/kafka-ui/3057
[ "connected" ]
f9906b5d307da11303177c4ddc04d4658c4ebfab
c2be45fd06263169feb514fe3bfe419bafcee377
[]
[]
"2022-12-09T22:29:39Z"
[ "type/bug", "scope/frontend", "status/accepted", "status/confirmed" ]
[UI] Offset & timestamp filters do not work
On the Messages page when Offset/Timestamp fields **are filled** we need to pass seekType argument with OFFSET/TIMESTAMP values correspondingly. When Offset/Timestamp are **not filled** we should pass seekType=LATEST if "Newest first" mode is on and seekType=BEGINNING when mode is "Oldest first".
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/BackwardRecordEmitter.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/ForwardRecordEmitter.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/TailingEmitter.java", "kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx" ]
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/BackwardRecordEmitter.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/ForwardRecordEmitter.java", "kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/TailingEmitter.java", "kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx" ]
[]
diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/BackwardRecordEmitter.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/BackwardRecordEmitter.java index c651bd5e56b..996f8b9f70c 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/BackwardRecordEmitter.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/BackwardRecordEmitter.java @@ -17,6 +17,7 @@ import org.apache.kafka.clients.consumer.ConsumerRecord; import org.apache.kafka.clients.consumer.KafkaConsumer; import org.apache.kafka.common.TopicPartition; +import org.apache.kafka.common.errors.InterruptException; import org.apache.kafka.common.utils.Bytes; import reactor.core.publisher.FluxSink; @@ -85,6 +86,9 @@ public void accept(FluxSink<TopicMessageEventDTO> sink) { } sendFinishStatsAndCompleteSink(sink); log.debug("Polling finished"); + } catch (InterruptException kafkaInterruptException) { + log.debug("Polling finished due to thread interruption"); + sink.complete(); } catch (Exception e) { log.error("Error occurred while consuming records", e); sink.error(e); diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/ForwardRecordEmitter.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/ForwardRecordEmitter.java index 132b4790297..9fadb149d48 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/ForwardRecordEmitter.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/ForwardRecordEmitter.java @@ -9,6 +9,7 @@ import org.apache.kafka.clients.consumer.ConsumerRecord; import org.apache.kafka.clients.consumer.ConsumerRecords; import org.apache.kafka.clients.consumer.KafkaConsumer; +import org.apache.kafka.common.errors.InterruptException; import org.apache.kafka.common.utils.Bytes; import reactor.core.publisher.FluxSink; @@ -59,6 +60,9 @@ public void accept(FluxSink<TopicMessageEventDTO> sink) { } sendFinishStatsAndCompleteSink(sink); log.debug("Polling finished"); + } catch (InterruptException kafkaInterruptException) { + log.debug("Polling finished due to thread interruption"); + sink.complete(); } catch (Exception e) { log.error("Error occurred while consuming records", e); sink.error(e); diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/TailingEmitter.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/TailingEmitter.java index 025dd874cfa..06cd8dad998 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/TailingEmitter.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/emitter/TailingEmitter.java @@ -41,6 +41,7 @@ public void accept(FluxSink<TopicMessageEventDTO> sink) { sink.complete(); log.debug("Tailing finished"); } catch (InterruptException kafkaInterruptException) { + log.debug("Tailing finished due to thread interruption"); sink.complete(); } catch (Exception e) { log.error("Error consuming {}", consumerPosition, e); diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx index 9f54e56d5a4..0e3d3d46614 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx @@ -219,6 +219,15 @@ const Filters: React.FC<FiltersProps> = ({ default: props.seekType = currentSeekType; } + + if (offset && currentSeekType === SeekType.OFFSET) { + props.seekType = SeekType.OFFSET; + } + + if (timestamp && currentSeekType === SeekType.TIMESTAMP) { + props.seekType = SeekType.TIMESTAMP; + } + props.seekTo = selectedPartitions.map(({ value }) => { const offsetProperty = seekDirection === SeekDirection.FORWARD ? 'offsetMin' : 'offsetMax';
null
train
val
2022-12-12T11:15:55
"2022-09-21T14:06:49Z"
iliax
train
provectus/kafka-ui/3037_3066
provectus/kafka-ui
provectus/kafka-ui/3037
provectus/kafka-ui/3066
[ "connected" ]
c13440b65d602399d27d7110580fe18856ad0603
de7f7db26919e7c23d49ea2c3e38f455205dec5e
[]
[ "why do we need to define variable? topic name is used only once, so let's specify it directly to openTopic(). fix it also at 104 line pls", "we already made methods like navigateToTopicsAndOpenDetails() and I've asked u to use them when the navigation is not the main scope of test", "does it have only @id='offset' or it depends on seekTypeDdl?", "1. add linebreak before forEach() pls\n2. ur assertion doesn't work because u missed isTrue() at the end\n3. u missed as() method to add assertion description\n4. offset instance as far as i know is always integer, that's why i made getOffset() return int. why did u set setOffsetMessagesField() to handle string? Integer.parseInt(offsetValue) looks kinky", "is there any //select element in DOM?", "change method to handle integer", "done", "Done", "refactored to $x(\"//label[text()='Seek Type']//..//input\");", "1. done\r\n2. done\r\n3. done\r\n4. done", "done\r\n", "nothing elements with select", "as we discussed, this field also can get date value, so we need to change int argument to String and set String.valueOf(seekTypeValue) from test", "done", "ok but what about method name?", "renamed to setSeekTypeValueFldMessagesTab()" ]
"2022-12-12T13:21:12Z"
[ "scope/QA", "status/accepted", "scope/AQA" ]
[e2e]Checking messages filtering by Offset within Topic/Messages
Autotest implementation for: https://app.qase.io/case/KAFKAUI-15 Description: Checking messages filtering by Offset within Messages/Topic Pre-conditions: - Login to Kafka-ui application - Open the 'Local' section - Select the 'Topics' - Open the Topic profile - Turn to Messages tab Steps: 1. Enter the value in Offset input field Input data: "1" 2. Click 'Submit' button Expected result: Messages which have the entered Offset should be filtered
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicMessagesTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java index 4a38e7909d7..6b328ab22e0 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java @@ -1,18 +1,20 @@ package com.provectus.kafka.ui.pages.topic; +import static com.codeborne.selenide.Selenide.$; +import static com.codeborne.selenide.Selenide.$$x; +import static com.codeborne.selenide.Selenide.$x; +import static org.apache.commons.lang.math.RandomUtils.nextInt; + import com.codeborne.selenide.CollectionCondition; import com.codeborne.selenide.Condition; import com.codeborne.selenide.ElementsCollection; import com.codeborne.selenide.SelenideElement; import com.provectus.kafka.ui.pages.BasePage; import io.qameta.allure.Step; -import org.openqa.selenium.By; import java.util.ArrayList; import java.util.Arrays; import java.util.List; - -import static com.codeborne.selenide.Selenide.*; -import static org.apache.commons.lang.math.RandomUtils.nextInt; +import org.openqa.selenium.By; public class TopicDetails extends BasePage { @@ -20,6 +22,8 @@ public class TopicDetails extends BasePage { protected SelenideElement messageAmountCell = $x("//tbody/tr/td[5]"); protected SelenideElement overviewTab = $x("//a[contains(text(),'Overview')]"); protected SelenideElement messagesTab = $x("//a[contains(text(),'Messages')]"); + protected SelenideElement seekTypeDdl = $x("//ul[@id='selectSeekType']/li"); + protected SelenideElement seekTypeField = $x("//label[text()='Seek Type']//..//input"); protected SelenideElement addFiltersBtn = $x("//button[text()='Add Filters']"); protected SelenideElement savedFiltersLink = $x("//div[text()='Saved Filters']"); protected SelenideElement addFilterCodeModalTitle = $x("//label[text()='Filter code']"); @@ -38,6 +42,7 @@ public class TopicDetails extends BasePage { protected SelenideElement partitionsField = $x("//div[contains(text(),'Partitions')]/../span"); protected SelenideElement backToCreateFiltersLink = $x("//div[text()='Back To create filters']"); protected ElementsCollection messageGridItems = $$x("//tbody//tr"); + protected String seekFilterDdlLocator = "//ul[@id='selectSeekType']/ul/li[text()='%s']"; protected String savedFilterNameLocator = "//div[@role='savedFilter']/div[contains(text(),'%s')]"; protected String consumerIdLocator = "//a[@title='%s']"; protected String topicHeaderLocator = "//h1[contains(text(),'%s')]"; @@ -120,6 +125,26 @@ public TopicDetails clickProduceMessageBtn() { return this; } + @Step + public TopicDetails selectSeekTypeDdlMessagesTab(String seekTypeName){ + seekTypeDdl.shouldBe(Condition.enabled).click(); + $x(String.format(seekFilterDdlLocator, seekTypeName)).shouldBe(Condition.visible).click(); + return this; + } + + @Step + public TopicDetails setSeekTypeValueFldMessagesTab(String seekTypeValue){ + seekTypeField.shouldBe(Condition.enabled).sendKeys(seekTypeValue); + return this; + } + + @Step + public TopicDetails clickSubmitFiltersBtnMessagesTab(){ + clickByJavaScript(submitBtn); + waitUntilSpinnerDisappear(); + return this; + } + @Step public TopicDetails clickMessagesAddFiltersBtn() { addFiltersBtn.shouldBe(Condition.enabled).click(); @@ -248,6 +273,11 @@ public TopicDetails.MessageGridItem getMessage(int offset) { .findFirst().orElse(null); } + @Step + public List<MessageGridItem> getAllMessages(){ + return initItems(); + } + @Step public TopicDetails.MessageGridItem getRandomMessage() { return getMessage(nextInt(initItems().size() - 1));
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicMessagesTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicMessagesTests.java index 8fbf915d543..67d531c8a67 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicMessagesTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicMessagesTests.java @@ -1,7 +1,7 @@ package com.provectus.kafka.ui.suite.topics; import static com.provectus.kafka.ui.pages.BasePage.AlertHeader.SUCCESS; -import static com.provectus.kafka.ui.pages.NaviSideBar.SideMenuOption.TOPICS; +import static com.provectus.kafka.ui.pages.topic.TopicDetails.TopicMenu.MESSAGES; import static com.provectus.kafka.ui.settings.Source.CLUSTER_NAME; import static com.provectus.kafka.ui.utilities.FileUtils.fileToString; import static org.apache.commons.lang.RandomStringUtils.randomAlphabetic; @@ -59,8 +59,10 @@ void produceMessage() { topicDetails .waitUntilScreenReady(); SoftAssertions softly = new SoftAssertions(); - softly.assertThat(topicDetails.isKeyMessageVisible((TOPIC_FOR_MESSAGES.getMessageKey()))).withFailMessage("isKeyMessageVisible()").isTrue(); - softly.assertThat(topicDetails.isContentMessageVisible((TOPIC_FOR_MESSAGES.getMessageContent()).trim())).withFailMessage("isContentMessageVisible()").isTrue(); + softly.assertThat(topicDetails.isKeyMessageVisible((TOPIC_FOR_MESSAGES.getMessageKey()))) + .withFailMessage("isKeyMessageVisible()").isTrue(); + softly.assertThat(topicDetails.isContentMessageVisible((TOPIC_FOR_MESSAGES.getMessageContent()).trim())) + .withFailMessage("isContentMessageVisible()").isTrue(); softly.assertAll(); } @@ -100,8 +102,7 @@ void clearMessage() { @CaseId(21) @Test void copyMessageFromTopicProfile() { - String topicName = "_schemas"; - navigateToTopicsAndOpenDetails(topicName); + navigateToTopicsAndOpenDetails("_schemas"); topicDetails .openDetailsTab(TopicDetails.TopicMenu.MESSAGES) .getRandomMessage() @@ -111,6 +112,28 @@ void copyMessageFromTopicProfile() { "isAlertWithMessageVisible()"); } + @Disabled + @Issue("https://github.com/provectus/kafka-ui/issues/2856") + @DisplayName("Checking messages filtering by Offset within Topic/Messages") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(15) + @Test + void checkingMessageFilteringByOffset() { + String offsetValue = "2"; + navigateToTopicsAndOpenDetails("_schemas"); + topicDetails + .openDetailsTab(MESSAGES) + .selectSeekTypeDdlMessagesTab("Offset") + .setSeekTypeValueFldMessagesTab(offsetValue) + .clickSubmitFiltersBtnMessagesTab(); + SoftAssertions softly = new SoftAssertions(); + topicDetails.getAllMessages() + .forEach(messages -> softly.assertThat(messages.getOffset() == Integer.parseInt(offsetValue)) + .as("getAllMessages()").isTrue()); + softly.assertAll(); + } + @AfterAll public void afterAll() { TOPIC_LIST.forEach(topic -> apiHelper.deleteTopic(CLUSTER_NAME, topic.getName()));
test
val
2022-12-16T13:32:20
"2022-12-07T10:56:11Z"
anezboretskiy
train
provectus/kafka-ui/2885_3067
provectus/kafka-ui
provectus/kafka-ui/2885
provectus/kafka-ui/3067
[ "connected" ]
c2be45fd06263169feb514fe3bfe419bafcee377
a0a464d455a7ee6fa02965206b00754fa1e89e49
[ "For context -- we had an outage where the topic config was edited, but the person did not notice that this field was showing value different from the one was actually intended for the topic (because they were editing a different parameter), therefore on \"save\" topic retention.ms was changed from 60s to 7d.", "Hey Alex,\r\n\r\ncould you check if that's still an issue on `master`-labeled image? I think we've done fixing something like this previously. \r\nThanks.", "@Haarolean Yes this appears to be fixed at [fce4c23](https://github.com/provectus/kafka-ui/commit/fce4c23276ebd798f182a2af7150cabffb2d63db).\r\n\r\nCan this please be backported to 0.4.x. We can't use unstable version in our production environment and current stable release can easily cause outage (also it's been 6 months since last release, and I guess the changes coming in new version are large, hence I don't know when is the new stable release going to land).", "@akamensky hey, I can build a new image just for you, since we don't have workflows to do a minor release branch.\r\nThe new release is gonna be out soon (once we're done with #753 frontend and testing)\r\n", "@Haarolean BTW, we found an issue with the fix. In the form it will show correct value (i.e. cleanup.policy == compact), but when submitting form it will submit \"default\" value there. It works if manually select correct value in field, if don't touch it even it shows correct value the one that goes to server may break things.", "@akamensky thanks, we'll investigate." ]
[]
"2022-12-12T14:04:24Z"
[ "type/bug", "scope/frontend", "status/accepted", "status/confirmed" ]
[FE] Fix cleanup policy submit/display
**Describe the bug** All of our topics have a custom retention.ms (usually 1h-10h). When using this UI to edit topic settings, the UI always shows cluster default for retention.ms (604800000), which we need to remember to set to correct value (even if this field not supposed to change). **Set up** 1. Running current stable release (0.4.0) as docker-compose service **Steps to Reproduce** Steps to reproduce the behavior: 1. Create topic with retention.ms different from cluster default 2. Go to edit topic settings 3. Check the retention.ms field value **Expected behavior** Expected behavior -- the field always show the retention.ms of the topic correctly. **Actual behavior** The field always show cluster default value. **Screenshots** ![Screenshot from 2022-11-07 14-05-32](https://user-images.githubusercontent.com/985805/200218070-2eabd56f-4ee2-4b44-aede-fd8386b7b021.png) ![Screenshot from 2022-11-07 14-05-14](https://user-images.githubusercontent.com/985805/200218073-3bb3417b-03c7-47cf-88de-f41d5cc126c6.png) **Note** After more testing, same behavior is for `max.message.bytes`, `min.insync.replicas` and `cleanup.policy`. Which makes it seems that all non "custom parameters" behave like this.
[ "kafka-ui-react-app/src/components/Topics/Topic/Edit/topicParamsTransformer.ts", "kafka-ui-react-app/src/components/Topics/shared/Form/TopicForm.tsx" ]
[ "kafka-ui-react-app/src/components/Topics/Topic/Edit/topicParamsTransformer.ts", "kafka-ui-react-app/src/components/Topics/shared/Form/TopicForm.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Edit/topicParamsTransformer.ts b/kafka-ui-react-app/src/components/Topics/Topic/Edit/topicParamsTransformer.ts index e60faf11ccb..910f31a5fd2 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Edit/topicParamsTransformer.ts +++ b/kafka-ui-react-app/src/components/Topics/Topic/Edit/topicParamsTransformer.ts @@ -4,6 +4,7 @@ import { TOPIC_CUSTOM_PARAMS_PREFIX, } from 'lib/constants'; import { TOPIC_EDIT_FORM_DEFAULT_PROPS } from 'components/Topics/Topic/Edit/Edit'; +import { getCleanUpPolicyValue } from 'components/Topics/shared/Form/TopicForm'; import { Topic, TopicConfig } from 'generated-sources'; export const getValue = ( @@ -30,6 +31,9 @@ const topicParamsTransformer = (topic?: Topic, config?: TopicConfig[]) => { replicationFactor: topic.replicationFactor, partitions: topic.partitionCount || TOPIC_EDIT_FORM_DEFAULT_PROPS.partitions, + cleanupPolicy: + getCleanUpPolicyValue(topic.cleanUpPolicy) || + TOPIC_EDIT_FORM_DEFAULT_PROPS.cleanupPolicy, maxMessageBytes: getValue(config, 'max.message.bytes', 1000012), minInSyncReplicas: getValue(config, 'min.insync.replicas', 1), retentionBytes: getValue(config, 'retention.bytes', -1), diff --git a/kafka-ui-react-app/src/components/Topics/shared/Form/TopicForm.tsx b/kafka-ui-react-app/src/components/Topics/shared/Form/TopicForm.tsx index a44aa266f77..15f60eaaac2 100644 --- a/kafka-ui-react-app/src/components/Topics/shared/Form/TopicForm.tsx +++ b/kafka-ui-react-app/src/components/Topics/shared/Form/TopicForm.tsx @@ -36,6 +36,17 @@ const CleanupPolicyOptions: Array<SelectOption> = [ { value: 'compact,delete', label: 'Compact,Delete' }, ]; +export const getCleanUpPolicyValue = (cleanUpPolicy?: string) => { + if (!cleanUpPolicy) return undefined; + + return CleanupPolicyOptions.find((option: SelectOption) => { + return ( + option.value.toString().replace(/,/g, '_') === + cleanUpPolicy?.toLowerCase() + ); + })?.value.toString(); +}; + const RetentionBytesOptions: Array<SelectOption> = [ { value: NOT_SET, label: 'Not Set' }, { value: BYTES_IN_GB, label: '1 GB' }, @@ -61,12 +72,7 @@ const TopicForm: React.FC<Props> = ({ const navigate = useNavigate(); const { clusterName } = useAppParams<{ clusterName: ClusterName }>(); const getCleanUpPolicy = - CleanupPolicyOptions.find((option: SelectOption) => { - return ( - option.value.toString().replace(/,/g, '_') === - cleanUpPolicy?.toLowerCase() - ); - })?.value || CleanupPolicyOptions[0].value; + getCleanUpPolicyValue(cleanUpPolicy) || CleanupPolicyOptions[0].value; const getRetentionBytes = RetentionBytesOptions.find((option: SelectOption) => { @@ -97,8 +103,8 @@ const TopicForm: React.FC<Props> = ({ </S.NameField> </S.Column> - {!isEditing && ( - <S.Column> + <S.Column> + {!isEditing && ( <div> <InputLabel htmlFor="topicFormNumberOfPartitions"> Number of partitions * @@ -114,32 +120,33 @@ const TopicForm: React.FC<Props> = ({ <ErrorMessage errors={errors} name="partitions" /> </FormError> </div> - <div> - <InputLabel - id="topicFormCleanupPolicyLabel" - htmlFor="topicFormCleanupPolicy" - > - Cleanup policy - </InputLabel> - <Controller - defaultValue={CleanupPolicyOptions[0].value} - control={control} - name="cleanupPolicy" - render={({ field: { name, onChange } }) => ( - <Select - id="topicFormCleanupPolicy" - aria-labelledby="topicFormCleanupPolicyLabel" - name={name} - value={getCleanUpPolicy} - onChange={onChange} - minWidth="250px" - options={CleanupPolicyOptions} - /> - )} - /> - </div> - </S.Column> - )} + )} + + <div> + <InputLabel + id="topicFormCleanupPolicyLabel" + htmlFor="topicFormCleanupPolicy" + > + Cleanup policy + </InputLabel> + <Controller + defaultValue={CleanupPolicyOptions[0].value} + control={control} + name="cleanupPolicy" + render={({ field: { name, onChange } }) => ( + <Select + id="topicFormCleanupPolicy" + aria-labelledby="topicFormCleanupPolicyLabel" + name={name} + value={getCleanUpPolicy} + onChange={onChange} + minWidth="250px" + options={CleanupPolicyOptions} + /> + )} + /> + </div> + </S.Column> </fieldset> <S.Column>
null
test
val
2022-12-19T18:44:14
"2022-11-07T03:06:08Z"
akamensky
train
provectus/kafka-ui/2778_3080
provectus/kafka-ui
provectus/kafka-ui/2778
provectus/kafka-ui/3080
[ "connected" ]
1117b296a7843272a8827099b9220aee4e5d9ab5
5cdd44daee610d32fcade22c11f929048ac1535e
[ "Hey, can i be assigned to this", "@Drex72 any luck?", "@Haarolean this can be checked by QA", "Bug still exists on master, let's make status ready after merging.\r\nWill test related PR https://github.com/provectus/kafka-ui/pull/3080" ]
[ "i think it is best to transform this `redux` functionality clear -> `react-query` , hence the `query` will handle delete as well, this way it is mixed.", "it is still un-removed here , we should change it like the upper one.", "i done it ", "`clearMessage` or `clearMessages`", "i change it ", "this whole file can be deleted as well, since it serves no use , `state.isFetching` is controlled by query hook, so that already gives `isLoading`.", "should i remove whole topicMessages reducer ?\r\nbecause if I delete TopicMessagesState, then it will be necessary to change a lot of files and tests" ]
"2022-12-14T12:38:36Z"
[ "type/bug", "good first issue", "scope/frontend", "status/accepted", "status/confirmed" ]
The Message count is not updating with Clearing message for partition
**Describe the bug** Message count stays the same after clearing message for partition in Topic's overview page. Needs refreshing the page. **Set up** https://www.kafka-ui.provectus.io/ **Steps to Reproduce** 1. Navigate to Topic's overview 2. Open the menu for partition 3. Press Clear Messages **Expected behavior** Messages count should be updated immediately **Screenshots** https://user-images.githubusercontent.com/104780608/196624761-0847ba39-4c30-4562-8890-31af9ec985c8.mov **Additional context** <!-- (Add any other context about the problem here) -->
[ "kafka-ui-react-app/src/components/Topics/List/ActionsCell.tsx", "kafka-ui-react-app/src/components/Topics/List/BatchActionsBar.tsx", "kafka-ui-react-app/src/components/Topics/List/__tests__/TopicTable.spec.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Overview/ActionsCell.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Overview/__test__/Overview.spec.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Topic.tsx", "kafka-ui-react-app/src/components/Topics/Topic/__test__/Topic.spec.tsx", "kafka-ui-react-app/src/lib/hooks/api/topics.ts", "kafka-ui-react-app/src/redux/reducers/topicMessages/__test__/reducer.spec.ts", "kafka-ui-react-app/src/redux/reducers/topicMessages/topicMessagesSlice.ts" ]
[ "kafka-ui-react-app/src/components/Topics/List/ActionsCell.tsx", "kafka-ui-react-app/src/components/Topics/List/BatchActionsBar.tsx", "kafka-ui-react-app/src/components/Topics/List/__tests__/TopicTable.spec.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Overview/ActionsCell.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Overview/__test__/Overview.spec.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Topic.tsx", "kafka-ui-react-app/src/components/Topics/Topic/__test__/Topic.spec.tsx", "kafka-ui-react-app/src/lib/hooks/api/topics.ts", "kafka-ui-react-app/src/redux/reducers/topicMessages/__test__/reducer.spec.ts", "kafka-ui-react-app/src/redux/reducers/topicMessages/topicMessagesSlice.ts" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/List/ActionsCell.tsx b/kafka-ui-react-app/src/components/Topics/List/ActionsCell.tsx index 4fb5f26c7d2..2f6fe3f49af 100644 --- a/kafka-ui-react-app/src/components/Topics/List/ActionsCell.tsx +++ b/kafka-ui-react-app/src/components/Topics/List/ActionsCell.tsx @@ -1,20 +1,17 @@ import React from 'react'; import { Action, CleanUpPolicy, Topic, ResourceType } from 'generated-sources'; import { CellContext } from '@tanstack/react-table'; -import { useAppDispatch } from 'lib/hooks/redux'; import ClusterContext from 'components/contexts/ClusterContext'; import { ClusterNameRoute } from 'lib/paths'; import useAppParams from 'lib/hooks/useAppParams'; -import { clearTopicMessages } from 'redux/reducers/topicMessages/topicMessagesSlice'; import { Dropdown, DropdownItem, DropdownItemHint, } from 'components/common/Dropdown'; -import { useQueryClient } from '@tanstack/react-query'; import { - topicKeys, useDeleteTopic, + useClearTopicMessages, useRecreateTopic, } from 'lib/hooks/api/topics'; import { ActionDropdownItem } from 'components/common/ActionComponent'; @@ -24,20 +21,16 @@ const ActionsCell: React.FC<CellContext<Topic, unknown>> = ({ row }) => { const { isReadOnly, isTopicDeletionAllowed } = React.useContext(ClusterContext); - const dispatch = useAppDispatch(); const { clusterName } = useAppParams<ClusterNameRoute>(); - const queryClient = useQueryClient(); + const clearMessages = useClearTopicMessages(clusterName); const deleteTopic = useDeleteTopic(clusterName); const recreateTopic = useRecreateTopic({ clusterName, topicName: name }); const disabled = internal || isReadOnly; const clearTopicMessagesHandler = async () => { - await dispatch( - clearTopicMessages({ clusterName, topicName: name }) - ).unwrap(); - return queryClient.invalidateQueries(topicKeys.all(clusterName)); + await clearMessages.mutateAsync(name); }; const isCleanupDisabled = cleanUpPolicy !== CleanUpPolicy.DELETE; diff --git a/kafka-ui-react-app/src/components/Topics/List/BatchActionsBar.tsx b/kafka-ui-react-app/src/components/Topics/List/BatchActionsBar.tsx index d75f4d0da2c..a04429be738 100644 --- a/kafka-ui-react-app/src/components/Topics/List/BatchActionsBar.tsx +++ b/kafka-ui-react-app/src/components/Topics/List/BatchActionsBar.tsx @@ -3,11 +3,13 @@ import { Row } from '@tanstack/react-table'; import { Action, Topic, ResourceType } from 'generated-sources'; import useAppParams from 'lib/hooks/useAppParams'; import { ClusterName } from 'redux/interfaces'; -import { topicKeys, useDeleteTopic } from 'lib/hooks/api/topics'; +import { + topicKeys, + useClearTopicMessages, + useDeleteTopic, +} from 'lib/hooks/api/topics'; import { useConfirm } from 'lib/hooks/useConfirm'; import { Button } from 'components/common/Button/Button'; -import { useAppDispatch } from 'lib/hooks/redux'; -import { clearTopicMessages } from 'redux/reducers/topicMessages/topicMessagesSlice'; import { clusterTopicCopyRelativePath } from 'lib/paths'; import { useQueryClient } from '@tanstack/react-query'; import { ActionCanButton } from 'components/common/ActionComponent'; @@ -25,11 +27,14 @@ const BatchActionsbar: React.FC<BatchActionsbarProps> = ({ }) => { const { clusterName } = useAppParams<{ clusterName: ClusterName }>(); const confirm = useConfirm(); - const dispatch = useAppDispatch(); const deleteTopic = useDeleteTopic(clusterName); const selectedTopics = rows.map(({ original }) => original.name); const client = useQueryClient(); + const clearMessages = useClearTopicMessages(clusterName); + const clearTopicMessagesHandler = async (topicName: Topic['name']) => { + await clearMessages.mutateAsync(topicName); + }; const deleteTopicsHandler = () => { confirm('Are you sure you want to remove selected topics?', async () => { try { @@ -50,7 +55,7 @@ const BatchActionsbar: React.FC<BatchActionsbarProps> = ({ try { await Promise.all( selectedTopics.map((topicName) => - dispatch(clearTopicMessages({ clusterName, topicName })).unwrap() + clearTopicMessagesHandler(topicName) ) ); resetRowSelection(); diff --git a/kafka-ui-react-app/src/components/Topics/List/__tests__/TopicTable.spec.tsx b/kafka-ui-react-app/src/components/Topics/List/__tests__/TopicTable.spec.tsx index 5e494cdfc8f..61be3f1360c 100644 --- a/kafka-ui-react-app/src/components/Topics/List/__tests__/TopicTable.spec.tsx +++ b/kafka-ui-react-app/src/components/Topics/List/__tests__/TopicTable.spec.tsx @@ -6,16 +6,15 @@ import { externalTopicPayload, topicsPayload } from 'lib/fixtures/topics'; import ClusterContext from 'components/contexts/ClusterContext'; import userEvent from '@testing-library/user-event'; import { + useClearTopicMessages, useDeleteTopic, useRecreateTopic, useTopics, } from 'lib/hooks/api/topics'; import TopicTable from 'components/Topics/List/TopicTable'; import { clusterTopicsPath } from 'lib/paths'; -import { useAppDispatch } from 'lib/hooks/redux'; const clusterName = 'test-cluster'; -const unwrapMock = jest.fn(); jest.mock('lib/hooks/redux', () => ({ ...jest.requireActual('lib/hooks/redux'), @@ -29,22 +28,24 @@ jest.mock('lib/hooks/api/topics', () => ({ useDeleteTopic: jest.fn(), useRecreateTopic: jest.fn(), useTopics: jest.fn(), + useClearTopicMessages: jest.fn(), })); const deleteTopicMock = jest.fn(); const recreateTopicMock = jest.fn(); +const clearTopicMessages = jest.fn(); describe('TopicTable Components', () => { beforeEach(() => { (useDeleteTopic as jest.Mock).mockImplementation(() => ({ mutateAsync: deleteTopicMock, })); + (useClearTopicMessages as jest.Mock).mockImplementation(() => ({ + mutateAsync: clearTopicMessages, + })); (useRecreateTopic as jest.Mock).mockImplementation(() => ({ mutateAsync: recreateTopicMock, })); - (useAppDispatch as jest.Mock).mockImplementation(() => () => ({ - unwrap: unwrapMock, - })); }); const renderComponent = ( @@ -185,9 +186,9 @@ describe('TopicTable Components', () => { ).toBeInTheDocument(); const confirmBtn = getButtonByName('Confirm'); expect(confirmBtn).toBeInTheDocument(); - expect(unwrapMock).not.toHaveBeenCalled(); + expect(clearTopicMessages).not.toHaveBeenCalled(); await userEvent.click(confirmBtn); - expect(unwrapMock).toHaveBeenCalledTimes(2); + expect(clearTopicMessages).toHaveBeenCalledTimes(2); expect(screen.getAllByRole('checkbox')[1]).not.toBeChecked(); expect(screen.getAllByRole('checkbox')[2]).not.toBeChecked(); }); @@ -282,7 +283,7 @@ describe('TopicTable Components', () => { await userEvent.click( screen.getByRole('button', { name: 'Confirm' }) ); - expect(unwrapMock).toHaveBeenCalled(); + expect(clearTopicMessages).toHaveBeenCalled(); }); }); diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Overview/ActionsCell.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Overview/ActionsCell.tsx index 0e1eed205a9..19e03a1411c 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Overview/ActionsCell.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Overview/ActionsCell.tsx @@ -1,13 +1,11 @@ import React from 'react'; import { Action, Partition, ResourceType } from 'generated-sources'; import { CellContext } from '@tanstack/react-table'; -import { useAppDispatch } from 'lib/hooks/redux'; import ClusterContext from 'components/contexts/ClusterContext'; import { RouteParamsClusterTopic } from 'lib/paths'; import useAppParams from 'lib/hooks/useAppParams'; -import { clearTopicMessages } from 'redux/reducers/topicMessages/topicMessagesSlice'; import { Dropdown } from 'components/common/Dropdown'; -import { useTopicDetails } from 'lib/hooks/api/topics'; +import { useClearTopicMessages, useTopicDetails } from 'lib/hooks/api/topics'; import { ActionDropdownItem } from 'components/common/ActionComponent'; const ActionsCell: React.FC<CellContext<Partition, unknown>> = ({ row }) => { @@ -15,12 +13,11 @@ const ActionsCell: React.FC<CellContext<Partition, unknown>> = ({ row }) => { const { data } = useTopicDetails({ clusterName, topicName }); const { isReadOnly } = React.useContext(ClusterContext); const { partition } = row.original; - const dispatch = useAppDispatch(); + + const clearMessages = useClearTopicMessages(clusterName, [partition]); const clearTopicMessagesHandler = async () => { - await dispatch( - clearTopicMessages({ clusterName, topicName, partitions: [partition] }) - ).unwrap(); + await clearMessages.mutateAsync(topicName); }; const disabled = data?.internal || isReadOnly || data?.cleanUpPolicy !== 'DELETE'; diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Overview/__test__/Overview.spec.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Overview/__test__/Overview.spec.tsx index da05b108c73..bf762a002b3 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Overview/__test__/Overview.spec.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Overview/__test__/Overview.spec.tsx @@ -8,8 +8,7 @@ import ClusterContext from 'components/contexts/ClusterContext'; import userEvent from '@testing-library/user-event'; import { clusterTopicPath } from 'lib/paths'; import { Replica } from 'components/Topics/Topic/Overview/Overview.styled'; -import { useTopicDetails } from 'lib/hooks/api/topics'; -import { useAppDispatch } from 'lib/hooks/redux'; +import { useClearTopicMessages, useTopicDetails } from 'lib/hooks/api/topics'; import { externalTopicPayload, internalTopicPayload, @@ -26,14 +25,10 @@ const defaultContextValues = { jest.mock('lib/hooks/api/topics', () => ({ useTopicDetails: jest.fn(), + useClearTopicMessages: jest.fn(), })); -const unwrapMock = jest.fn(); - -jest.mock('lib/hooks/redux', () => ({ - ...jest.requireActual('lib/hooks/redux'), - useAppDispatch: jest.fn(), -})); +const clearTopicMessage = jest.fn(); describe('Overview', () => { const renderComponent = ( @@ -43,6 +38,9 @@ describe('Overview', () => { (useTopicDetails as jest.Mock).mockImplementation(() => ({ data: topic, })); + (useClearTopicMessages as jest.Mock).mockImplementation(() => ({ + mutateAsync: clearTopicMessage, + })); const path = clusterTopicPath(clusterName, topicName); return render( <WithRoute path={clusterTopicPath()}> @@ -54,12 +52,6 @@ describe('Overview', () => { ); }; - beforeEach(() => { - (useAppDispatch as jest.Mock).mockImplementation(() => () => ({ - unwrap: unwrapMock, - })); - }); - it('at least one replica was rendered', () => { renderComponent(); expect(screen.getByLabelText('replica-info')).toBeInTheDocument(); @@ -136,7 +128,7 @@ describe('Overview', () => { const clearMessagesButton = screen.getByText('Clear Messages'); await userEvent.click(clearMessagesButton); - expect(unwrapMock).toHaveBeenCalledTimes(1); + expect(clearTopicMessage).toHaveBeenCalledTimes(1); }); }); diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Topic.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Topic.tsx index 01e140c8bb0..8945523576b 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Topic.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Topic.tsx @@ -21,14 +21,12 @@ import { useAppDispatch } from 'lib/hooks/redux'; import useAppParams from 'lib/hooks/useAppParams'; import { Dropdown, DropdownItemHint } from 'components/common/Dropdown'; import { + useClearTopicMessages, useDeleteTopic, useRecreateTopic, useTopicDetails, } from 'lib/hooks/api/topics'; -import { - clearTopicMessages, - resetTopicMessages, -} from 'redux/reducers/topicMessages/topicMessagesSlice'; +import { resetTopicMessages } from 'redux/reducers/topicMessages/topicMessagesSlice'; import { Action, CleanUpPolicy, ResourceType } from 'generated-sources'; import PageLoader from 'components/common/PageLoader/PageLoader'; import SlidingSidebar from 'components/common/SlidingSidebar'; @@ -69,9 +67,11 @@ const Topic: React.FC = () => { dispatch(resetTopicMessages()); }; }, []); - + const clearMessages = useClearTopicMessages(clusterName); + const clearTopicMessagesHandler = async () => { + await clearMessages.mutateAsync(topicName); + }; const canCleanup = data?.cleanUpPolicy === CleanUpPolicy.DELETE; - return ( <> <PageHeading @@ -110,9 +110,7 @@ const Topic: React.FC = () => { </ActionDropdownItem> <ActionDropdownItem - onClick={() => - dispatch(clearTopicMessages({ clusterName, topicName })).unwrap() - } + onClick={clearTopicMessagesHandler} confirm="Are you sure want to clear topic messages?" disabled={!canCleanup} danger diff --git a/kafka-ui-react-app/src/components/Topics/Topic/__test__/Topic.spec.tsx b/kafka-ui-react-app/src/components/Topics/Topic/__test__/Topic.spec.tsx index a686b38e2ba..460e4ad5dee 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/__test__/Topic.spec.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/__test__/Topic.spec.tsx @@ -16,6 +16,7 @@ import { import { CleanUpPolicy, Topic } from 'generated-sources'; import { externalTopicPayload } from 'lib/fixtures/topics'; import { + useClearTopicMessages, useDeleteTopic, useRecreateTopic, useTopicDetails, @@ -31,9 +32,11 @@ jest.mock('lib/hooks/api/topics', () => ({ useTopicDetails: jest.fn(), useDeleteTopic: jest.fn(), useRecreateTopic: jest.fn(), + useClearTopicMessages: jest.fn(), })); const unwrapMock = jest.fn(); +const clearTopicMessages = jest.fn(); jest.mock('lib/hooks/redux', () => ({ ...jest.requireActual('lib/hooks/redux'), @@ -98,6 +101,9 @@ describe('Details', () => { (useRecreateTopic as jest.Mock).mockImplementation(() => ({ mutateAsync: mockRecreate, })); + (useClearTopicMessages as jest.Mock).mockImplementation(() => ({ + mutateAsync: clearTopicMessages, + })); (useAppDispatch as jest.Mock).mockImplementation(() => () => ({ unwrap: unwrapMock, })); @@ -145,7 +151,7 @@ describe('Details', () => { name: 'Confirm', })[0]; await waitFor(() => userEvent.click(submitButton)); - expect(unwrapMock).toHaveBeenCalledTimes(1); + expect(clearTopicMessages).toHaveBeenCalledTimes(1); }); it('closes the modal when cancel button is clicked', async () => { diff --git a/kafka-ui-react-app/src/lib/hooks/api/topics.ts b/kafka-ui-react-app/src/lib/hooks/api/topics.ts index 9af0b63d620..23e6d65ce37 100644 --- a/kafka-ui-react-app/src/lib/hooks/api/topics.ts +++ b/kafka-ui-react-app/src/lib/hooks/api/topics.ts @@ -2,6 +2,7 @@ import { topicsApiClient as api, messagesApiClient as messagesApi, consumerGroupsApiClient, + messagesApiClient, } from 'lib/api'; import { useMutation, useQuery, useQueryClient } from '@tanstack/react-query'; import { @@ -233,6 +234,34 @@ export function useDeleteTopic(clusterName: ClusterName) { } ); } + +export function useClearTopicMessages( + clusterName: ClusterName, + partitions?: number[] +) { + const client = useQueryClient(); + return useMutation( + async (topicName: Topic['name']) => { + await messagesApiClient.deleteTopicMessages({ + clusterName, + partitions, + topicName, + }); + return topicName; + }, + + { + onSuccess: (topicName) => { + showSuccessAlert({ + id: `message-${topicName}-${clusterName}-${partitions}`, + message: `${topicName} messages have been successfully cleared!`, + }); + client.invalidateQueries(topicKeys.all(clusterName)); + }, + } + ); +} + export function useRecreateTopic(props: GetTopicDetailsRequest) { const client = useQueryClient(); return useMutation(() => api.recreateTopic(props), { diff --git a/kafka-ui-react-app/src/redux/reducers/topicMessages/__test__/reducer.spec.ts b/kafka-ui-react-app/src/redux/reducers/topicMessages/__test__/reducer.spec.ts index c2cfa73f273..b9fa4cd4ec1 100644 --- a/kafka-ui-react-app/src/redux/reducers/topicMessages/__test__/reducer.spec.ts +++ b/kafka-ui-react-app/src/redux/reducers/topicMessages/__test__/reducer.spec.ts @@ -1,6 +1,5 @@ import reducer, { addTopicMessage, - clearTopicMessages, resetTopicMessages, updateTopicMessagesMeta, updateTopicMessagesPhase, @@ -12,9 +11,6 @@ import { topicMessagesMetaPayload, } from './fixtures'; -const clusterName = 'local'; -const topicName = 'localTopic'; - describe('TopicMessages reducer', () => { it('Adds new message', () => { const state = reducer( @@ -67,24 +63,6 @@ describe('TopicMessages reducer', () => { expect(newState.messages.length).toEqual(0); }); - it('clear messages', () => { - const state = reducer( - undefined, - addTopicMessage({ message: topicMessagePayload }) - ); - expect(state.messages.length).toEqual(1); - - expect( - reducer(state, { - type: clearTopicMessages.fulfilled, - payload: { clusterName, topicName }, - }) - ).toEqual({ - ...state, - messages: [], - }); - }); - it('Updates Topic Messages Phase', () => { const phase = 'Polling'; diff --git a/kafka-ui-react-app/src/redux/reducers/topicMessages/topicMessagesSlice.ts b/kafka-ui-react-app/src/redux/reducers/topicMessages/topicMessagesSlice.ts index 5de97980f31..846cbd0c978 100644 --- a/kafka-ui-react-app/src/redux/reducers/topicMessages/topicMessagesSlice.ts +++ b/kafka-ui-react-app/src/redux/reducers/topicMessages/topicMessagesSlice.ts @@ -1,36 +1,6 @@ -import { createAsyncThunk, createSlice } from '@reduxjs/toolkit'; -import { TopicMessagesState, ClusterName, TopicName } from 'redux/interfaces'; +import { createSlice } from '@reduxjs/toolkit'; +import { TopicMessagesState } from 'redux/interfaces'; import { TopicMessage } from 'generated-sources'; -import { - getResponse, - showServerError, - showSuccessAlert, -} from 'lib/errorHandling'; -import { messagesApiClient } from 'lib/api'; - -export const clearTopicMessages = createAsyncThunk< - undefined, - { clusterName: ClusterName; topicName: TopicName; partitions?: number[] } ->( - 'topicMessages/clearTopicMessages', - async ({ clusterName, topicName, partitions }, { rejectWithValue }) => { - try { - await messagesApiClient.deleteTopicMessages({ - clusterName, - topicName, - partitions, - }); - showSuccessAlert({ - id: `message-${topicName}-${clusterName}-${partitions}`, - message: `${topicName} messages have been successfully cleared!`, - }); - return undefined; - } catch (err) { - showServerError(err as Response); - return rejectWithValue(await getResponse(err as Response)); - } - } -); export const initialState: TopicMessagesState = { messages: [], @@ -68,11 +38,6 @@ const topicMessagesSlice = createSlice({ state.isFetching = action.payload; }, }, - extraReducers: (builder) => { - builder.addCase(clearTopicMessages.fulfilled, (state) => { - state.messages = []; - }); - }, }); export const {
null
train
val
2023-03-14T19:03:05
"2022-10-19T07:28:03Z"
armenuikafka
train
provectus/kafka-ui/2974_3081
provectus/kafka-ui
provectus/kafka-ui/2974
provectus/kafka-ui/3081
[ "connected" ]
c148f112a404815d6645fa97209199eced054728
696cde7dccd655e656c19c613bf54240e47ffded
[]
[ "do we have an `enum` for this ? `'Fallback'`", "i didn't find ", "Move it to styled.ts", "I would suggest to create ellipsis common component with width property ", "Color should be moved to theme", "Color should be moved to theme", "done", "done", "done", "done", "```suggestion\r\nconst Ellipsis: React.FC<PropsWithChildren<{ text: string }>> = ({ text, children }) => {\r\n return (\r\n <S.Wrapper>\r\n <S.Text>{children}</S.Text>\r\n {children && (<S.Postfix>{children}</S.Postfix>)}\r\n </S.Wrapper>\r\n );\r\n};\r\n```", "```suggestion\r\nexport const Wrapper = styled.div`\r\n display: flex;\r\n gap: 8px;\r\n align-items: center;\r\n`;\r\n\r\n```", "Ellipsis is a common component. why we sending serde property here?", "```suggestion\r\n <Ellipsis Π΅ΡƒΡ‡Π΅={renderFilteredJson(key, keyFilters)}>\r\n {serde === 'Fallback' && <WarningRedIcon />}\r\n </Ellipsis>\r\n```", "```suggestion\r\n <S.DataCell title={key}>\r\n```", "It must be color from `Colors` object", "i think something stayed in russian here", "done", "done", "done", "done", "done", "done" ]
"2022-12-14T18:31:50Z"
[ "type/enhancement", "good first issue", "scope/frontend", "status/accepted" ]
FE: Mark serde-failed messages with red
mark red in case of keySerde / valueSerde == β€œFallback” https://www.figma.com/file/ZkvysdRUmt9p2PDNU3a9WU/Kafka-UI?node-id=3161%3A16359&t=Hq9IHVUrCgjb9fqt-0
[ "kafka-ui-react-app/src/components/Topics/Topic/Messages/Message.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/MessageContent/MessageContent.styled.ts", "kafka-ui-react-app/src/theme/theme.ts" ]
[ "kafka-ui-react-app/src/components/Topics/Topic/Messages/Message.tsx", "kafka-ui-react-app/src/components/Topics/Topic/Messages/MessageContent/MessageContent.styled.ts", "kafka-ui-react-app/src/components/common/Ellipsis/Ellipsis.styled.ts", "kafka-ui-react-app/src/components/common/Ellipsis/Ellipsis.tsx", "kafka-ui-react-app/src/components/common/Icons/WarningRedIcon.tsx", "kafka-ui-react-app/src/theme/theme.ts" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Message.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Message.tsx index 60f09b82933..0282cde2ea7 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Message.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Message.tsx @@ -1,5 +1,4 @@ import React from 'react'; -import styled from 'styled-components'; import useDataSaver from 'lib/hooks/useDataSaver'; import { TopicMessage } from 'generated-sources'; import MessageToggleIcon from 'components/common/Icons/MessageToggleIcon'; @@ -7,22 +6,12 @@ import IconButtonWrapper from 'components/common/Icons/IconButtonWrapper'; import { Dropdown, DropdownItem } from 'components/common/Dropdown'; import { formatTimestamp } from 'lib/dateTimeHelpers'; import { JSONPath } from 'jsonpath-plus'; +import Ellipsis from 'components/common/Ellipsis/Ellipsis'; +import WarningRedIcon from 'components/common/Icons/WarningRedIcon'; import MessageContent from './MessageContent/MessageContent'; import * as S from './MessageContent/MessageContent.styled'; -const StyledDataCell = styled.td` - overflow: hidden; - white-space: nowrap; - text-overflow: ellipsis; - max-width: 350px; - min-width: 350px; -`; - -const ClickableRow = styled.tr` - cursor: pointer; -`; - export interface PreviewFilter { field: string; path: string; @@ -43,6 +32,8 @@ const Message: React.FC<Props> = ({ partition, content, headers, + valueSerde, + keySerde, }, keyFilters, contentFilters, @@ -100,7 +91,7 @@ const Message: React.FC<Props> = ({ return ( <> - <ClickableRow + <S.ClickableRow onMouseEnter={() => setVEllipsisOpen(true)} onMouseLeave={() => setVEllipsisOpen(false)} onClick={toggleIsOpen} @@ -115,16 +106,20 @@ const Message: React.FC<Props> = ({ <td> <div>{formatTimestamp(timestamp)}</div> </td> - <StyledDataCell title={key}> - {renderFilteredJson(key, keyFilters)} - </StyledDataCell> - <StyledDataCell title={content}> + <S.DataCell title={key}> + <Ellipsis text={renderFilteredJson(key, keyFilters)}> + {keySerde === 'Fallback' && <WarningRedIcon />} + </Ellipsis> + </S.DataCell> + <S.DataCell title={content}> <S.Metadata> <S.MetadataValue> - {renderFilteredJson(content, contentFilters)} + <Ellipsis text={renderFilteredJson(content, contentFilters)}> + {valueSerde === 'Fallback' && <WarningRedIcon />} + </Ellipsis> </S.MetadataValue> </S.Metadata> - </StyledDataCell> + </S.DataCell> <td style={{ width: '5%' }}> {vEllipsisOpen && ( <Dropdown> @@ -135,7 +130,7 @@ const Message: React.FC<Props> = ({ </Dropdown> )} </td> - </ClickableRow> + </S.ClickableRow> {isOpen && ( <MessageContent messageKey={key} diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessageContent/MessageContent.styled.ts b/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessageContent/MessageContent.styled.ts index 9eb5e6b0627..1d8672042b7 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessageContent/MessageContent.styled.ts +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/MessageContent/MessageContent.styled.ts @@ -35,7 +35,16 @@ export const ContentBox = styled.div` flex-grow: 1; } `; - +export const DataCell = styled.td` + overflow: hidden; + white-space: nowrap; + text-overflow: ellipsis; + max-width: 350px; + min-width: 350px; +`; +export const ClickableRow = styled.tr` + cursor: pointer; +`; export const MetadataWrapper = styled.div` background-color: ${({ theme }) => theme.topicMetaData.backgroundColor}; padding: 24px; diff --git a/kafka-ui-react-app/src/components/common/Ellipsis/Ellipsis.styled.ts b/kafka-ui-react-app/src/components/common/Ellipsis/Ellipsis.styled.ts new file mode 100644 index 00000000000..d301090b46c --- /dev/null +++ b/kafka-ui-react-app/src/components/common/Ellipsis/Ellipsis.styled.ts @@ -0,0 +1,14 @@ +import styled from 'styled-components'; + +export const Text = styled.div` + overflow: hidden; + white-space: nowrap; + text-overflow: ellipsis; + max-width: 340px; +`; + +export const Wrapper = styled.div` + display: flex; + gap: 8px; + align-items: center; +`; diff --git a/kafka-ui-react-app/src/components/common/Ellipsis/Ellipsis.tsx b/kafka-ui-react-app/src/components/common/Ellipsis/Ellipsis.tsx new file mode 100644 index 00000000000..f6a690d9c60 --- /dev/null +++ b/kafka-ui-react-app/src/components/common/Ellipsis/Ellipsis.tsx @@ -0,0 +1,20 @@ +import React, { PropsWithChildren } from 'react'; + +import * as S from './Ellipsis.styled'; + +type EllipsisProps = { + text: React.ReactNode; +}; + +const Ellipsis: React.FC<PropsWithChildren<EllipsisProps>> = ({ + text, + children, +}) => { + return ( + <S.Wrapper> + <S.Text>{text}</S.Text> + {children} + </S.Wrapper> + ); +}; +export default Ellipsis; diff --git a/kafka-ui-react-app/src/components/common/Icons/WarningRedIcon.tsx b/kafka-ui-react-app/src/components/common/Icons/WarningRedIcon.tsx new file mode 100644 index 00000000000..13231f58944 --- /dev/null +++ b/kafka-ui-react-app/src/components/common/Icons/WarningRedIcon.tsx @@ -0,0 +1,32 @@ +import React from 'react'; +import { useTheme } from 'styled-components'; + +const WarningRedIcon: React.FC = () => { + const theme = useTheme(); + return ( + <svg + width="20" + height="20" + viewBox="0 0 20 20" + fill="none" + xmlns="http://www.w3.org/2000/svg" + > + <rect + width="20" + height="20" + rx="10" + fill={theme.icons.warningRedIcon.rectFill} + /> + <path + d="M9 4.74219H11V12.7422H9V4.74219Z" + fill={theme.icons.warningRedIcon.pathFill} + /> + <path + d="M9 14.7422C9 14.1899 9.44772 13.7422 10 13.7422C10.5523 13.7422 11 14.1899 11 14.7422C11 15.2945 10.5523 15.7422 10 15.7422C9.44772 15.7422 9 15.2945 9 14.7422Z" + fill={theme.icons.warningRedIcon.pathFill} + /> + </svg> + ); +}; + +export default WarningRedIcon; diff --git a/kafka-ui-react-app/src/theme/theme.ts b/kafka-ui-react-app/src/theme/theme.ts index 978b694913f..33dbf1c619e 100644 --- a/kafka-ui-react-app/src/theme/theme.ts +++ b/kafka-ui-react-app/src/theme/theme.ts @@ -173,6 +173,10 @@ const baseTheme = { closeIcon: Colors.neutral[30], deleteIcon: Colors.red[20], warningIcon: Colors.yellow[20], + warningRedIcon: { + rectFill: Colors.red[10], + pathFill: Colors.red[50], + }, messageToggleIcon: { normal: Colors.brand[30], hover: Colors.brand[40],
null
train
val
2023-04-13T09:57:40
"2022-11-25T10:26:42Z"
Haarolean
train
provectus/kafka-ui/2538_3089
provectus/kafka-ui
provectus/kafka-ui/2538
provectus/kafka-ui/3089
[ "connected" ]
de21721e00b5109d6792630707db6cff568df824
1bcdec4accc7ef9e319a3f671183f01eb4e5bdba
[ "`/api/clusters/{clusterName}/consumer-groups/paged` -- no topicname\r\n`/api/clusters/<cluster>/topics/<topicName>/consumer-groups` -- no query", "@Haarolean Can you please assign this to me ?", "@shubhwip yo, any plans to do this?:)", "@Haarolean I will do it probably this weekend or next." ]
[ "i suggest writing it like this , since we are not using `getData` function anywhere else \r\n\r\n`const consumerGroups = React.useMemo(\r\n () => data.filter((item) => item.groupId.toLocaleLowerCase().indexOf(keyword) > -1),\r\n [data, keyword]\r\n );`\r\n" ]
"2022-12-18T17:07:20Z"
[ "type/enhancement", "good first issue", "scope/frontend", "status/accepted" ]
Search consumers by name within topic details
**Describe the bug** Implement the Search within Topic/Consumers tab **Set up** https://www.kafka-ui.provectus.io/ **Steps to Reproduce** 1. Navigate to Topics 2. Open the Topic 3. Turn to Consumers tab **Expected behavior** Search by **consumer Name** should be available within Topic/Consumers tab **Screenshots** <img width="1717" alt="search topic:consumers" src="https://user-images.githubusercontent.com/104780608/188839052-d4e88426-4a19-4d93-aa56-6574a94e53ba.png">
[ "kafka-ui-react-app/src/components/Topics/Topic/ConsumerGroups/TopicConsumerGroups.tsx" ]
[ "kafka-ui-react-app/src/components/Topics/Topic/ConsumerGroups/TopicConsumerGroups.styled.ts", "kafka-ui-react-app/src/components/Topics/Topic/ConsumerGroups/TopicConsumerGroups.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/Topic/ConsumerGroups/TopicConsumerGroups.styled.ts b/kafka-ui-react-app/src/components/Topics/Topic/ConsumerGroups/TopicConsumerGroups.styled.ts new file mode 100644 index 00000000000..77768dc9980 --- /dev/null +++ b/kafka-ui-react-app/src/components/Topics/Topic/ConsumerGroups/TopicConsumerGroups.styled.ts @@ -0,0 +1,6 @@ +import styled from 'styled-components'; + +export const SearchWrapper = styled.div` + margin: 10px; + width: 21%; +`; diff --git a/kafka-ui-react-app/src/components/Topics/Topic/ConsumerGroups/TopicConsumerGroups.tsx b/kafka-ui-react-app/src/components/Topics/Topic/ConsumerGroups/TopicConsumerGroups.tsx index 69948fd66d4..173c28c76e3 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/ConsumerGroups/TopicConsumerGroups.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/ConsumerGroups/TopicConsumerGroups.tsx @@ -5,14 +5,27 @@ import useAppParams from 'lib/hooks/useAppParams'; import { useTopicConsumerGroups } from 'lib/hooks/api/topics'; import { ColumnDef } from '@tanstack/react-table'; import Table, { LinkCell, TagCell } from 'components/common/NewTable'; +import Search from 'components/common/Search/Search'; + +import * as S from './TopicConsumerGroups.styled'; const TopicConsumerGroups: React.FC = () => { + const [keyword, setKeyword] = React.useState(''); const { clusterName, topicName } = useAppParams<RouteParamsClusterTopic>(); - const { data: consumerGroups = [] } = useTopicConsumerGroups({ + const { data = [] } = useTopicConsumerGroups({ clusterName, topicName, }); + + const consumerGroups = React.useMemo( + () => + data.filter( + (item) => item.groupId.toLocaleLowerCase().indexOf(keyword) > -1 + ), + [data, keyword] + ); + const columns = React.useMemo<ColumnDef<ConsumerGroup>[]>( () => [ { @@ -61,12 +74,21 @@ const TopicConsumerGroups: React.FC = () => { [] ); return ( - <Table - columns={columns} - data={consumerGroups} - enableSorting - emptyMessage="No active consumer groups" - /> + <> + <S.SearchWrapper> + <Search + onChange={setKeyword} + placeholder="Search by Consumer Name" + value={keyword} + /> + </S.SearchWrapper> + <Table + columns={columns} + data={consumerGroups} + enableSorting + emptyMessage="No active consumer groups" + /> + </> ); };
null
val
val
2023-04-03T07:10:14
"2022-09-07T09:08:53Z"
armenuikafka
train
provectus/kafka-ui/3079_3094
provectus/kafka-ui
provectus/kafka-ui/3079
provectus/kafka-ui/3094
[ "connected" ]
f8417e6eaaf29f1fe0dbbb09da55dc4e701020ec
ca3916aae01494a890ba021535400ff148df0fcb
[]
[ "openBrokersTab -> openDetailsTab", "Condition.visible?", "extra space", "in previous implementation we checked that getAllVisibleElements() list is also enabled. seems here this assertion is missed", "u already have this waiting at 44 line", "page wasn't changed, so why do we need waitUntilScreenReady() again? ", "seems u've missed define here new screen brokersConfigTabPanel", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "There was extra assertion.", "what about 37, 69, 70, 71 lines?", "getBrokersTabs -> getDetailsTabs", "BrokerMenu -> DetailsTab", "Stream.of(\"Log directories\", \"Configs\", \"Metrics\") -> DetailsTab.values()", "Fixed.", "Fixed.", "Fixed.", "Fixed." ]
"2022-12-19T13:30:11Z"
[ "scope/QA", "scope/AQA" ]
[e2e] Update case "Checking the information of used Broker in a cluster"
Autotest implementation for: https://app.qase.io/case/KAFKAUI-85 Needs to add two steps : 2)Click on Broker 3) Click on Configs tab Expected result : 2)The Broker's profile should open with following information: Segment Size Segment Count Port Host Log directories tab: Name Error Topics Partitions Enabled Configs tab Enabled Metrics tab 3) Should be displayed configs list with following information: Key, Value (with edit functionality), Source and Search by Key input field
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersList.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersConfigTabPanel.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersDetails.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersList.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java index d7b342b0254..cf559af1e2c 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/BasePage.java @@ -21,7 +21,7 @@ public abstract class BasePage extends WebUtils { protected ElementsCollection allGridItems = $$x("//tr[@class]"); protected String summaryCellLocator = "//div[contains(text(),'%s')]"; protected String tableElementNameLocator = "//tbody//a[contains(text(),'%s')]"; - protected String columnHeaderLocator = "//table//tr/th/div[text()='%s']"; + protected String columnHeaderLocator = "//table//tr/th//div[text()='%s']"; protected void waitUntilSpinnerDisappear() { log.debug("\nwaitUntilSpinnerDisappear"); diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersConfigTabPanel.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersConfigTabPanel.java new file mode 100644 index 00000000000..b7f1402a57e --- /dev/null +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersConfigTabPanel.java @@ -0,0 +1,40 @@ +package com.provectus.kafka.ui.pages.brokers; + +import static com.codeborne.selenide.Selenide.$$x; +import static com.codeborne.selenide.Selenide.$x; + +import com.codeborne.selenide.Condition; +import com.codeborne.selenide.SelenideElement; +import com.provectus.kafka.ui.pages.BasePage; +import io.qameta.allure.Step; +import java.util.List; +import java.util.stream.Collectors; +import java.util.stream.Stream; + +public class BrokersConfigTabPanel extends BasePage { + + protected List<SelenideElement> editBtn = $$x("//button[@aria-label='editAction']"); + protected SelenideElement searchByKeyField = $x("//input[@placeholder='Search by Key']"); + + @Step + public BrokersConfigTabPanel waitUntilScreenReady(){ + waitUntilSpinnerDisappear(); + searchByKeyField.shouldBe(Condition.visible); + return this; + } + + @Step + public boolean isSearchByKeyVisible() { + return isVisible(searchByKeyField); + } + + public List<SelenideElement> getColumnHeaders() { + return Stream.of("Key", "Value", "Source") + .map(name -> $x(String.format(columnHeaderLocator, name))) + .collect(Collectors.toList()); + } + + public List<SelenideElement> getEditButtons() { + return editBtn; + } +} diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersDetails.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersDetails.java index 77fc9e0b41c..4eca65f1f47 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersDetails.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersDetails.java @@ -1,5 +1,6 @@ package com.provectus.kafka.ui.pages.brokers; +import static com.codeborne.selenide.Selenide.$; import static com.codeborne.selenide.Selenide.$x; import com.codeborne.selenide.Condition; @@ -11,11 +12,13 @@ import java.util.List; import java.util.stream.Collectors; import java.util.stream.Stream; +import org.openqa.selenium.By; public class BrokersDetails extends BasePage { protected SelenideElement logDirectoriesTab = $x("//a[text()='Log directories']"); protected SelenideElement metricsTab = $x("//a[text()='Metrics']"); + protected String brokersTabLocator = "//a[text()='%s']"; @Step public BrokersDetails waitUntilScreenReady() { @@ -24,6 +27,13 @@ public BrokersDetails waitUntilScreenReady() { return this; } + @Step + public BrokersDetails openDetailsTab(DetailsTab menu) { + $(By.linkText(menu.toString())).shouldBe(Condition.enabled).click(); + waitUntilSpinnerDisappear(); + return this; + } + private List<SelenideElement> getVisibleColumnHeaders() { return Stream.of("Name", "Topics", "Error", "Partitions") .map(name -> $x(String.format(columnHeaderLocator, name))) @@ -42,15 +52,40 @@ private List<SelenideElement> getVisibleSummaryCells() { .collect(Collectors.toList()); } + private List<SelenideElement> getDetailsTabs() { + return Stream.of(DetailsTab.values()) + .map(name -> $x(String.format(brokersTabLocator, name))) + .collect(Collectors.toList()); + } + @Step public List<SelenideElement> getAllEnabledElements() { - return getEnabledColumnHeaders(); + List<SelenideElement> enabledElements = new ArrayList<>(getEnabledColumnHeaders()); + enabledElements.addAll(getDetailsTabs()); + return enabledElements; } @Step public List<SelenideElement> getAllVisibleElements() { List<SelenideElement> visibleElements = new ArrayList<>(getVisibleSummaryCells()); visibleElements.addAll(getVisibleColumnHeaders()); + visibleElements.addAll(getDetailsTabs()); return visibleElements; } + + public enum DetailsTab { + LOG_DIRECTORIES("Log directories"), + CONFIGS("Configs"), + METRICS("Metrics"); + + private final String value; + + DetailsTab(String value) { + this.value = value; + } + + public String toString() { + return value; + } + } } diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersList.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersList.java index 98ed497e422..239a6359343 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersList.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersList.java @@ -24,7 +24,7 @@ public BrokersList waitUntilScreenReady() { } @Step - public BrokersList openBroker(String brokerId) { + public BrokersList openBroker(int brokerId) { getBrokerItem(brokerId).openItem(); return this; } @@ -67,9 +67,9 @@ private List<BrokersList.BrokerGridItem> initGridItems() { } @Step - public BrokerGridItem getBrokerItem(String id){ + public BrokerGridItem getBrokerItem(int id){ return initGridItems().stream() - .filter(e ->e.getId().equals(id)) + .filter(e ->e.getId() == id) .findFirst().orElse(null); } @@ -91,8 +91,8 @@ private SelenideElement getIdElm() { } @Step - public String getId() { - return getIdElm().getText().trim(); + public int getId() { + return Integer.parseInt(getIdElm().getText().trim()); } @Step
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java index 279704b8b95..1e3ea4d630f 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java @@ -3,6 +3,7 @@ import com.provectus.kafka.ui.helpers.ApiHelper; import com.provectus.kafka.ui.pages.NaviSideBar; import com.provectus.kafka.ui.pages.TopPanel; +import com.provectus.kafka.ui.pages.brokers.BrokersConfigTabPanel; import com.provectus.kafka.ui.pages.brokers.BrokersDetails; import com.provectus.kafka.ui.pages.brokers.BrokersList; import com.provectus.kafka.ui.pages.connector.ConnectorCreateForm; @@ -36,4 +37,5 @@ public abstract class Facade { protected TopPanel topPanel = new TopPanel(); protected BrokersList brokersList = new BrokersList(); protected BrokersDetails brokersDetails = new BrokersDetails(); + protected BrokersConfigTabPanel brokersConfigTabPanel = new BrokersConfigTabPanel(); } diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java index 0b54a4015ce..2fc2adce35c 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java @@ -1,6 +1,7 @@ package com.provectus.kafka.ui.suite.brokers; import static com.provectus.kafka.ui.pages.NaviSideBar.SideMenuOption.BROKERS; +import static com.provectus.kafka.ui.pages.brokers.BrokersDetails.DetailsTab.CONFIGS; import static org.assertj.core.api.Assertions.assertThat; import com.codeborne.selenide.Condition; @@ -38,11 +39,18 @@ public void checkExistingBrokersInCluster(){ navigateToBrokers(); assertThat(brokersList.getAllBrokers()).as("getAllBrokers()").size().isGreaterThan(0); brokersList - .openBroker("1"); + .openBroker(1); brokersDetails .waitUntilScreenReady(); verifyElementsCondition(brokersDetails.getAllVisibleElements(), Condition.visible); - verifyElementsCondition(brokersDetails.getAllVisibleElements(), Condition.enabled); + verifyElementsCondition(brokersDetails.getAllEnabledElements(), Condition.enabled); + brokersDetails + .openDetailsTab(CONFIGS); + brokersConfigTabPanel + .waitUntilScreenReady(); + verifyElementsCondition(brokersConfigTabPanel.getColumnHeaders(), Condition.visible); + verifyElementsCondition(brokersConfigTabPanel.getEditButtons(), Condition.enabled); + assertThat(brokersConfigTabPanel.isSearchByKeyVisible()).as("isSearchByKeyVisible()").isTrue(); } @Step
train
val
2022-12-19T19:30:51
"2022-12-14T12:19:46Z"
ArthurNiedial
train
provectus/kafka-ui/2372_3106
provectus/kafka-ui
provectus/kafka-ui/2372
provectus/kafka-ui/3106
[ "connected" ]
eef63466fb3aee13a63d8071b1abf3de425f3f2b
9d9892752454e0522968ce225175fc1e6d03c322
[ "https://vitejs.dev/guide/build.html#public-base-path\r\n\r\nwhen can configure this during build, by this parameter , if we can actually manage to pass the parameter during build.\r\n\r\nbut this is not on the fly url change solution , this won't cut it.\r\n", "https://github.com/vitejs/vite/issues/3522" ]
[ "πŸ‘ " ]
"2022-12-20T12:41:57Z"
[ "type/enhancement", "scope/frontend", "status/accepted" ]
Implement a full support for custom context path
https://webpack.js.org/guides/public-path/
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/controller/StaticController.java", "kafka-ui-react-app/index.html", "kafka-ui-react-app/src/components/Cluster/Cluster.tsx", "kafka-ui-react-app/src/components/Cluster/__tests__/Cluster.spec.tsx", "kafka-ui-react-app/vite.config.ts" ]
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/controller/StaticController.java", "kafka-ui-react-app/index.html", "kafka-ui-react-app/src/components/Cluster/Cluster.tsx", "kafka-ui-react-app/src/components/Cluster/__tests__/Cluster.spec.tsx", "kafka-ui-react-app/vite.config.ts" ]
[]
diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/controller/StaticController.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/controller/StaticController.java index 696f47073b7..72138c8010c 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/controller/StaticController.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/controller/StaticController.java @@ -53,9 +53,7 @@ public String getRenderedFile(ServerWebExchange exchange, AtomicReference<String @SneakyThrows private String buildFile(Resource file, String contextPath) { return ResourceUtil.readAsString(file) - .replace("\"/assets/", "\"" + contextPath + "/assets/") - .replace("\"/favicon/", "\"" + contextPath + "/favicon/") - .replace("/manifest.json", contextPath + "/manifest.json") - .replace("window.basePath = ''", "window.basePath=\"" + contextPath + "\""); + .replace("\"assets/", "\"" + contextPath + "/assets/") + .replace("PUBLIC-PATH-VARIABLE", contextPath); } } diff --git a/kafka-ui-react-app/index.html b/kafka-ui-react-app/index.html index 7b4ed2772aa..aa3f89f8f90 100644 --- a/kafka-ui-react-app/index.html +++ b/kafka-ui-react-app/index.html @@ -3,7 +3,6 @@ <head> <meta charset="utf-8" /> <meta name="viewport" content="width=device-width, initial-scale=1" /> - <!-- Google fonts --> <link rel="preconnect" href="https://fonts.googleapis.com" /> <link rel="preconnect" href="https://fonts.gstatic.com" crossorigin /> @@ -13,14 +12,18 @@ /> <!-- Favicons --> - <link rel="icon" href="/favicon/favicon.ico" sizes="any" /> - <link rel="icon" href="/favicon/icon.svg" type="image/svg+xml" /> - <link rel="apple-touch-icon" href="/favicon/apple-touch-icon.png" /> - <link rel="manifest" href="/manifest.json" /> + <link rel="icon" href="PUBLIC-PATH-VARIABLE/favicon/favicon.ico" sizes="any" /> + <link rel="icon" href="PUBLIC-PATH-VARIABLE/favicon/icon.svg" type="image/svg+xml" /> + <link rel="apple-touch-icon" href="PUBLIC-PATH-VARIABLE/favicon/apple-touch-icon.png" /> + <link rel="manifest" href="PUBLIC-PATH-VARIABLE/manifest.json" /> <title>UI for Apache Kafka</title> <script type="text/javascript"> - window.basePath = ''; + window.basePath = 'PUBLIC-PATH-VARIABLE'; + + window.__assetsPathBuilder = function (importer) { + return window.basePath+ "/" + importer; + }; </script> </head> diff --git a/kafka-ui-react-app/src/components/Cluster/Cluster.tsx b/kafka-ui-react-app/src/components/Cluster/Cluster.tsx index 11cc975b4ff..89c4931754c 100644 --- a/kafka-ui-react-app/src/components/Cluster/Cluster.tsx +++ b/kafka-ui-react-app/src/components/Cluster/Cluster.tsx @@ -16,23 +16,15 @@ import { import ClusterContext from 'components/contexts/ClusterContext'; import PageLoader from 'components/common/PageLoader/PageLoader'; import { useClusters } from 'lib/hooks/api/clusters'; -import Brokers from 'components/Brokers/Brokers'; -import Topics from 'components/Topics/Topics'; -import Schemas from 'components/Schemas/Schemas'; -import Connect from 'components/Connect/Connect'; -import KsqlDb from 'components/KsqlDb/KsqlDb'; -import ConsumerGroups from 'components/ConsumerGroups/ConsumerGroups'; -// We can't use Lazy loading till we have a better way to update publicPath in runtime -// Now java app replaces paths in builded index.html file. -// const Brokers = React.lazy(() => import('components/Brokers/Brokers')); -// const Topics = React.lazy(() => import('components/Topics/Topics')); -// const Schemas = React.lazy(() => import('components/Schemas/Schemas')); -// const Connect = React.lazy(() => import('components/Connect/Connect')); -// const KsqlDb = React.lazy(() => import('components/KsqlDb/KsqlDb')); -// const ConsumerGroups = React.lazy( -// () => import('components/ConsumerGroups/ConsumerGroups') -// ); +const Brokers = React.lazy(() => import('components/Brokers/Brokers')); +const Topics = React.lazy(() => import('components/Topics/Topics')); +const Schemas = React.lazy(() => import('components/Schemas/Schemas')); +const Connect = React.lazy(() => import('components/Connect/Connect')); +const KsqlDb = React.lazy(() => import('components/KsqlDb/KsqlDb')); +const ConsumerGroups = React.lazy( + () => import('components/ConsumerGroups/ConsumerGroups') +); const Cluster: React.FC = () => { const { clusterName } = useAppParams<ClusterNameRoute>(); @@ -59,49 +51,51 @@ const Cluster: React.FC = () => { return ( <Suspense fallback={<PageLoader />}> <ClusterContext.Provider value={contextValue}> - <Routes> - <Route - path={getNonExactPath(clusterBrokerRelativePath)} - element={<Brokers />} - /> - <Route - path={getNonExactPath(clusterTopicsRelativePath)} - element={<Topics />} - /> - <Route - path={getNonExactPath(clusterConsumerGroupsRelativePath)} - element={<ConsumerGroups />} - /> - {contextValue.hasSchemaRegistryConfigured && ( + <Suspense fallback={<PageLoader />}> + <Routes> <Route - path={getNonExactPath(clusterSchemasRelativePath)} - element={<Schemas />} + path={getNonExactPath(clusterBrokerRelativePath)} + element={<Brokers />} /> - )} - {contextValue.hasKafkaConnectConfigured && ( <Route - path={getNonExactPath(clusterConnectsRelativePath)} - element={<Connect />} + path={getNonExactPath(clusterTopicsRelativePath)} + element={<Topics />} /> - )} - {contextValue.hasKafkaConnectConfigured && ( <Route - path={getNonExactPath(clusterConnectorsRelativePath)} - element={<Connect />} + path={getNonExactPath(clusterConsumerGroupsRelativePath)} + element={<ConsumerGroups />} /> - )} - {contextValue.hasKsqlDbConfigured && ( + {contextValue.hasSchemaRegistryConfigured && ( + <Route + path={getNonExactPath(clusterSchemasRelativePath)} + element={<Schemas />} + /> + )} + {contextValue.hasKafkaConnectConfigured && ( + <Route + path={getNonExactPath(clusterConnectsRelativePath)} + element={<Connect />} + /> + )} + {contextValue.hasKafkaConnectConfigured && ( + <Route + path={getNonExactPath(clusterConnectorsRelativePath)} + element={<Connect />} + /> + )} + {contextValue.hasKsqlDbConfigured && ( + <Route + path={getNonExactPath(clusterKsqlDbRelativePath)} + element={<KsqlDb />} + /> + )} <Route - path={getNonExactPath(clusterKsqlDbRelativePath)} - element={<KsqlDb />} + path="/" + element={<Navigate to={clusterBrokerRelativePath} replace />} /> - )} - <Route - path="/" - element={<Navigate to={clusterBrokerRelativePath} replace />} - /> - </Routes> - <Outlet /> + </Routes> + <Outlet /> + </Suspense> </ClusterContext.Provider> </Suspense> ); diff --git a/kafka-ui-react-app/src/components/Cluster/__tests__/Cluster.spec.tsx b/kafka-ui-react-app/src/components/Cluster/__tests__/Cluster.spec.tsx index 9fcb77a79a9..d7b2c6d2bb2 100644 --- a/kafka-ui-react-app/src/components/Cluster/__tests__/Cluster.spec.tsx +++ b/kafka-ui-react-app/src/components/Cluster/__tests__/Cluster.spec.tsx @@ -1,7 +1,7 @@ import React from 'react'; import { Cluster, ClusterFeaturesEnum } from 'generated-sources'; import ClusterComponent from 'components/Cluster/Cluster'; -import { screen } from '@testing-library/react'; +import { screen, waitFor } from '@testing-library/react'; import { render, WithRoute } from 'lib/testHelpers'; import { clusterBrokersPath, @@ -59,6 +59,9 @@ describe('Cluster', () => { </WithRoute>, { initialEntries: [pathname] } ); + await waitFor(() => { + expect(screen.queryByRole('progressbar')).not.toBeInTheDocument(); + }); }; it('renders Brokers', async () => { diff --git a/kafka-ui-react-app/vite.config.ts b/kafka-ui-react-app/vite.config.ts index 7b329779071..3320bb8cd87 100644 --- a/kafka-ui-react-app/vite.config.ts +++ b/kafka-ui-react-app/vite.config.ts @@ -18,6 +18,26 @@ export default defineConfig(({ mode }) => { build: { outDir: 'build', }, + experimental: { + renderBuiltUrl( + filename: string, + { + hostType, + }: { + hostId: string; + hostType: 'js' | 'css' | 'html'; + type: 'asset' | 'public'; + } + ) { + if (hostType === 'js') { + return { + runtime: `window.__assetsPathBuilder(${JSON.stringify(filename)})`, + }; + } + + return filename; + }, + }, define: { 'process.env.NODE_ENV': `"${mode}"`, 'process.env.VITE_TAG': `"${process.env.VITE_TAG}"`,
null
train
val
2023-01-03T11:52:28
"2022-08-01T16:49:03Z"
Haarolean
train
provectus/kafka-ui/2706_3116
provectus/kafka-ui
provectus/kafka-ui/2706
provectus/kafka-ui/3116
[ "connected" ]
cf9b653e46e42f82ded7d6833285f4e9467a2c6f
cf4571b964d2276b5385b7c46d71612c2296c925
[ "Hey, @Haarolean I will be working on this issue. I think there is some problem with the middleware. Request you to assign me. Thank You!", "@spike-spiegel-21 sure, go ahead :)", "Hey, @Haarolean Currently I am able to run the [kafka-ui-react-app](https://github.com/provectus/kafka-ui/tree/master/kafka-ui-react-app) but I don't know how to create consumer group. Could you please help me?", "@spike-spiegel-21 \r\nyou can install any KC connector which does consume messages.\r\nHere's a bunch of [examples](https://github.com/provectus/kafka-ui/tree/master/documentation/compose/connectors), e.g. s3-sink will consume messages and post them to AWS S3, you can configure that one for example.\r\nHere's a [compose file](https://github.com/provectus/kafka-ui/blob/master/documentation/compose/e2e-tests.yaml) which does install connectors for you.", "@spike-spiegel-21 any luck?", "Hey @Haarolean Sorry for being, was busy with some university stuff. I haven't tried it since prev comment. I will surely complete this task in the upcoming week. Thank you.", "Hello @Haarolean I was trying to run [kafka-ui.yaml](https://github.com/provectus/kafka-ui/blob/master/documentation/compose/kafka-ui.yaml) with `sudo docker-compose -f docker-compose.yml up` and encounterd these warnings and errors:\r\n\r\n1. `kafka0 | ERROR: Did you forget the update_run.sh file that came with this docker-compose.yml file?\r\nkafka1 | ERROR: Did you forget the update_run.sh file that came with this docker-compose.yml file?\r\nkafka0 exited with code 1\r\nkafka1 exited with code 1\r\n`\r\n2. `WARN org.apache.kafka.clients.ClientUtils - Couldn't resolve server kafka1:29092 from bootstrap.servers as DNS resolution failed for kafka1\r\n`\r\nCould you please help me resolve this, or should I create custom docker-compose file?", "@spike-spiegel-21 which OS are you on? Did you run the docker-compose from the project directory or created your own? Do you have `update_run.sh` available at the path specified in volume binds list?" ]
[]
"2022-12-21T13:31:26Z"
[ "type/bug", "good first issue", "scope/frontend", "status/accepted", "status/confirmed" ]
Deleting a consumer hangs the page
ERROR: type should be string, got "\r\nhttps://user-images.githubusercontent.com/1494347/194282581-2387ac9f-bee4-4d09-a42a-3fdd312ffb2d.mov\r\n\r\n"
[ "kafka-ui-react-app/src/components/ConsumerGroups/Details/Details.tsx", "kafka-ui-react-app/src/redux/reducers/consumerGroups/consumerGroupsSlice.ts" ]
[ "kafka-ui-react-app/src/components/ConsumerGroups/Details/Details.tsx", "kafka-ui-react-app/src/redux/reducers/consumerGroups/consumerGroupsSlice.ts" ]
[]
diff --git a/kafka-ui-react-app/src/components/ConsumerGroups/Details/Details.tsx b/kafka-ui-react-app/src/components/ConsumerGroups/Details/Details.tsx index 961410ffd2f..098a487081a 100644 --- a/kafka-ui-react-app/src/components/ConsumerGroups/Details/Details.tsx +++ b/kafka-ui-react-app/src/components/ConsumerGroups/Details/Details.tsx @@ -18,10 +18,9 @@ import TableHeaderCell from 'components/common/table/TableHeaderCell/TableHeader import { useAppDispatch, useAppSelector } from 'lib/hooks/redux'; import { deleteConsumerGroup, + selectById, fetchConsumerGroupDetails, getAreConsumerGroupDetailsFulfilled, - getIsConsumerGroupDeleted, - selectById, } from 'redux/reducers/consumerGroups/consumerGroupsSlice'; import getTagColor from 'components/common/Tag/getTagColor'; import { Dropdown } from 'components/common/Dropdown'; @@ -41,23 +40,19 @@ const Details: React.FC = () => { const consumerGroup = useAppSelector((state) => selectById(state, consumerGroupID) ); - const isDeleted = useAppSelector(getIsConsumerGroupDeleted); const isFetched = useAppSelector(getAreConsumerGroupDetailsFulfilled); React.useEffect(() => { dispatch(fetchConsumerGroupDetails({ clusterName, consumerGroupID })); }, [clusterName, consumerGroupID, dispatch]); - const onDelete = () => { - dispatch(deleteConsumerGroup({ clusterName, consumerGroupID })); + const onDelete = async () => { + const res = await dispatch( + deleteConsumerGroup({ clusterName, consumerGroupID }) + ).unwrap(); + if (res) navigate('../'); }; - React.useEffect(() => { - if (isDeleted) { - navigate('../'); - } - }, [clusterName, navigate, isDeleted]); - const onResetOffsets = () => { navigate(clusterConsumerGroupResetRelativePath); }; diff --git a/kafka-ui-react-app/src/redux/reducers/consumerGroups/consumerGroupsSlice.ts b/kafka-ui-react-app/src/redux/reducers/consumerGroups/consumerGroupsSlice.ts index 2b7709b0099..91420f04043 100644 --- a/kafka-ui-react-app/src/redux/reducers/consumerGroups/consumerGroupsSlice.ts +++ b/kafka-ui-react-app/src/redux/reducers/consumerGroups/consumerGroupsSlice.ts @@ -195,11 +195,6 @@ export const getAreConsumerGroupsPagedFulfilled = createSelector( (status) => status === AsyncRequestStatus.fulfilled ); -export const getIsConsumerGroupDeleted = createSelector( - createFetchingSelector('consumerGroups/deleteConsumerGroup'), - (status) => status === AsyncRequestStatus.fulfilled -); - export const getAreConsumerGroupDetailsFulfilled = createSelector( createFetchingSelector('consumerGroups/fetchConsumerGroupDetails'), (status) => status === AsyncRequestStatus.fulfilled
null
train
val
2023-01-09T13:39:24
"2022-10-06T09:49:47Z"
Haarolean
train
provectus/kafka-ui/3107_3118
provectus/kafka-ui
provectus/kafka-ui/3107
provectus/kafka-ui/3118
[ "connected" ]
ca3916aae01494a890ba021535400ff148df0fcb
a9d7d6c5153ea63999f69f0b0433362a0b5a9c74
[]
[ "u already added this topic at 406 line", "seems here we need topicSettingsTab.waitUntilScreenReady();", "can't we openDotMenu from here?", "extra indent", "seems we can set it to private w/o @Step, because we'll have only two methods using current inside this class", "Removed.", "Fixed.", "Fixed.", "Fixed.", "Fixed.", "not sure", "not sure", "Fixed for sure." ]
"2022-12-22T10:36:47Z"
[ "scope/QA", "scope/AQA" ]
[e2e] Checking Topics settings to make sure retention.bytes is right according to Max size on disk in GB selected value
Autotest implementation for: https://app.qase.io/case/KAFKAUI-56 Description: The purpose of this case is to make sure that retention.bytes value in Settings is right according to Max size on disk in GB fields' selected value (in case of 'Not set' value, retention.bytes should be '-1') Pre-conditions : Not set Post-conditions : Not set
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersConfigTabPanel.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersConfigTab.java", "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicSettingsTab.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java", "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersConfigTabPanel.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersConfigTab.java similarity index 90% rename from kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersConfigTabPanel.java rename to kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersConfigTab.java index b7f1402a57e..70bba6047a8 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersConfigTabPanel.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/brokers/BrokersConfigTab.java @@ -11,13 +11,13 @@ import java.util.stream.Collectors; import java.util.stream.Stream; -public class BrokersConfigTabPanel extends BasePage { +public class BrokersConfigTab extends BasePage { protected List<SelenideElement> editBtn = $$x("//button[@aria-label='editAction']"); protected SelenideElement searchByKeyField = $x("//input[@placeholder='Search by Key']"); @Step - public BrokersConfigTabPanel waitUntilScreenReady(){ + public BrokersConfigTab waitUntilScreenReady(){ waitUntilSpinnerDisappear(); searchByKeyField.shouldBe(Condition.visible); return this; diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicSettingsTab.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicSettingsTab.java new file mode 100644 index 00000000000..4bf78fb2af0 --- /dev/null +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicSettingsTab.java @@ -0,0 +1,65 @@ +package com.provectus.kafka.ui.pages.topic; + +import static com.codeborne.selenide.Selenide.$x; + +import com.codeborne.selenide.CollectionCondition; +import com.codeborne.selenide.Condition; +import com.codeborne.selenide.SelenideElement; +import com.provectus.kafka.ui.pages.BasePage; +import io.qameta.allure.Step; +import java.util.ArrayList; +import java.util.List; + +public class TopicSettingsTab extends BasePage { + + protected SelenideElement defaultValueColumnHeaderLocator = $x("//div[text() = 'Default Value']"); + + @Step + public TopicSettingsTab waitUntilScreenReady(){ + waitUntilSpinnerDisappear(); + defaultValueColumnHeaderLocator.shouldBe(Condition.visible); + return this; + } + + private List<SettingsGridItem> initGridItems() { + List<SettingsGridItem> gridItemList = new ArrayList<>(); + allGridItems.shouldHave(CollectionCondition.sizeGreaterThan(0)) + .forEach(item -> gridItemList.add(new SettingsGridItem(item))); + return gridItemList; + } + + private TopicSettingsTab.SettingsGridItem getItemByKey(String key){ + return initGridItems().stream() + .filter(e ->e.getKey().equals(key)) + .findFirst().orElse(null); + } + + @Step + public String getValueByKey(String key){ + return getItemByKey(key).getValue(); + } + + public static class SettingsGridItem extends BasePage { + + private final SelenideElement element; + + public SettingsGridItem(SelenideElement element) { + this.element = element; + } + + @Step + public String getKey(){ + return element.$x("./td[1]/span").getText().trim(); + } + + @Step + public String getValue(){ + return element.$x("./td[2]/span").getText().trim(); + } + + @Step + public String getDefaultValue() { + return element.$x("./td[3]/span").getText().trim(); + } + } +}
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java index 1e3ea4d630f..c39df88dc25 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/base/Facade.java @@ -3,7 +3,7 @@ import com.provectus.kafka.ui.helpers.ApiHelper; import com.provectus.kafka.ui.pages.NaviSideBar; import com.provectus.kafka.ui.pages.TopPanel; -import com.provectus.kafka.ui.pages.brokers.BrokersConfigTabPanel; +import com.provectus.kafka.ui.pages.brokers.BrokersConfigTab; import com.provectus.kafka.ui.pages.brokers.BrokersDetails; import com.provectus.kafka.ui.pages.brokers.BrokersList; import com.provectus.kafka.ui.pages.connector.ConnectorCreateForm; @@ -15,6 +15,7 @@ import com.provectus.kafka.ui.pages.schema.SchemaDetails; import com.provectus.kafka.ui.pages.schema.SchemaRegistryList; import com.provectus.kafka.ui.pages.topic.ProduceMessagePanel; +import com.provectus.kafka.ui.pages.topic.TopicSettingsTab; import com.provectus.kafka.ui.pages.topic.TopicCreateEditForm; import com.provectus.kafka.ui.pages.topic.TopicDetails; import com.provectus.kafka.ui.pages.topic.TopicsList; @@ -37,5 +38,6 @@ public abstract class Facade { protected TopPanel topPanel = new TopPanel(); protected BrokersList brokersList = new BrokersList(); protected BrokersDetails brokersDetails = new BrokersDetails(); - protected BrokersConfigTabPanel brokersConfigTabPanel = new BrokersConfigTabPanel(); + protected BrokersConfigTab brokersConfigTab = new BrokersConfigTab(); + protected TopicSettingsTab topicSettingsTab = new TopicSettingsTab(); } diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java index 2fc2adce35c..0cba13812c3 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/brokers/BrokersTests.java @@ -46,11 +46,11 @@ public void checkExistingBrokersInCluster(){ verifyElementsCondition(brokersDetails.getAllEnabledElements(), Condition.enabled); brokersDetails .openDetailsTab(CONFIGS); - brokersConfigTabPanel + brokersConfigTab .waitUntilScreenReady(); - verifyElementsCondition(brokersConfigTabPanel.getColumnHeaders(), Condition.visible); - verifyElementsCondition(brokersConfigTabPanel.getEditButtons(), Condition.enabled); - assertThat(brokersConfigTabPanel.isSearchByKeyVisible()).as("isSearchByKeyVisible()").isTrue(); + verifyElementsCondition(brokersConfigTab.getColumnHeaders(), Condition.visible); + verifyElementsCondition(brokersConfigTab.getEditButtons(), Condition.enabled); + assertThat(brokersConfigTab.isSearchByKeyVisible()).as("isSearchByKeyVisible()").isTrue(); } @Step diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java index e75c39f7e97..4eca498532f 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java @@ -5,6 +5,8 @@ import static com.provectus.kafka.ui.pages.topic.enums.CleanupPolicyValue.COMPACT; import static com.provectus.kafka.ui.pages.topic.enums.CleanupPolicyValue.DELETE; import static com.provectus.kafka.ui.pages.topic.enums.CustomParameterType.COMPRESSION_TYPE; +import static com.provectus.kafka.ui.pages.topic.enums.MaxSizeOnDisk.NOT_SET; +import static com.provectus.kafka.ui.pages.topic.enums.MaxSizeOnDisk.SIZE_1_GB; import static com.provectus.kafka.ui.pages.topic.enums.MaxSizeOnDisk.SIZE_20_GB; import static com.provectus.kafka.ui.settings.Source.CLUSTER_NAME; import static com.provectus.kafka.ui.utilities.FileUtils.fileToString; @@ -46,21 +48,26 @@ public class TopicsTests extends BaseTest { .setCustomParameterType(COMPRESSION_TYPE) .setCustomParameterValue("producer") .setCleanupPolicyValue(DELETE); - private static final Topic TOPIC_FOR_UPDATE = new Topic() + private static final Topic TOPIC_TO_UPDATE = new Topic() .setName("topic-to-update-" + randomAlphabetic(5)) + .setNumberOfPartitions(1) .setCleanupPolicyValue(COMPACT) .setTimeToRetainData("604800001") .setMaxSizeOnDisk(SIZE_20_GB) .setMaxMessageBytes("1000020") .setMessageKey(fileToString(System.getProperty("user.dir") + "/src/test/resources/producedkey.txt")) .setMessageContent(fileToString(System.getProperty("user.dir") + "/src/test/resources/testData.txt")); - + private static final Topic TOPIC_TO_CHECK_SETTINGS = new Topic() + .setName("new-topic-" + randomAlphabetic(5)) + .setNumberOfPartitions(1) + .setMaxMessageBytes("1000012") + .setMaxSizeOnDisk(NOT_SET); private static final Topic TOPIC_FOR_DELETE = new Topic().setName("topic-to-delete-" + randomAlphabetic(5)); private static final List<Topic> TOPIC_LIST = new ArrayList<>(); @BeforeAll public void beforeAll() { - TOPIC_LIST.addAll(List.of(TOPIC_FOR_UPDATE, TOPIC_FOR_DELETE)); + TOPIC_LIST.addAll(List.of(TOPIC_TO_UPDATE, TOPIC_FOR_DELETE)); TOPIC_LIST.forEach(topic -> apiHelper.createTopic(CLUSTER_NAME, topic.getName())); } @@ -121,33 +128,33 @@ void checkAvailableOperations() { @Test @Order(3) public void updateTopic() { - navigateToTopicsAndOpenDetails(TOPIC_FOR_UPDATE.getName()); + navigateToTopicsAndOpenDetails(TOPIC_TO_UPDATE.getName()); topicDetails .openDotMenu() .clickEditSettingsMenu(); topicCreateEditForm .waitUntilScreenReady() - .selectCleanupPolicy((TOPIC_FOR_UPDATE.getCleanupPolicyValue())) + .selectCleanupPolicy((TOPIC_TO_UPDATE.getCleanupPolicyValue())) .setMinInsyncReplicas(10) - .setTimeToRetainDataInMs(TOPIC_FOR_UPDATE.getTimeToRetainData()) - .setMaxSizeOnDiskInGB(TOPIC_FOR_UPDATE.getMaxSizeOnDisk()) - .setMaxMessageBytes(TOPIC_FOR_UPDATE.getMaxMessageBytes()) + .setTimeToRetainDataInMs(TOPIC_TO_UPDATE.getTimeToRetainData()) + .setMaxSizeOnDiskInGB(TOPIC_TO_UPDATE.getMaxSizeOnDisk()) + .setMaxMessageBytes(TOPIC_TO_UPDATE.getMaxMessageBytes()) .clickCreateTopicBtn(); topicDetails .waitUntilScreenReady(); - navigateToTopicsAndOpenDetails(TOPIC_FOR_UPDATE.getName()); + navigateToTopicsAndOpenDetails(TOPIC_TO_UPDATE.getName()); topicDetails .openDotMenu() .clickEditSettingsMenu(); SoftAssertions softly = new SoftAssertions(); softly.assertThat(topicCreateEditForm.getCleanupPolicy()).as("getCleanupPolicy()") - .isEqualTo(TOPIC_FOR_UPDATE.getCleanupPolicyValue().getVisibleText()); + .isEqualTo(TOPIC_TO_UPDATE.getCleanupPolicyValue().getVisibleText()); softly.assertThat(topicCreateEditForm.getTimeToRetain()).as("getTimeToRetain()") - .isEqualTo(TOPIC_FOR_UPDATE.getTimeToRetainData()); + .isEqualTo(TOPIC_TO_UPDATE.getTimeToRetainData()); softly.assertThat(topicCreateEditForm.getMaxSizeOnDisk()).as("getMaxSizeOnDisk()") - .isEqualTo(TOPIC_FOR_UPDATE.getMaxSizeOnDisk().getVisibleText()); + .isEqualTo(TOPIC_TO_UPDATE.getMaxSizeOnDisk().getVisibleText()); softly.assertThat(topicCreateEditForm.getMaxMessageBytes()).as("getMaxMessageBytes()") - .isEqualTo(TOPIC_FOR_UPDATE.getMaxMessageBytes()); + .isEqualTo(TOPIC_TO_UPDATE.getMaxMessageBytes()); softly.assertAll(); } @@ -381,6 +388,57 @@ void checkShowInternalTopicsButtonFunctionality(){ softly.assertAll(); } + @DisplayName("Checking Topics settings to make sure retention.bytes is right according to Max size on disk in GB selected value") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(56) + @Test + void checkRetentionBytesAccordingToMaxSizeOnDisk(){ + navigateToTopics(); + topicsList + .clickAddTopicBtn(); + topicCreateEditForm + .waitUntilScreenReady() + .setTopicName(TOPIC_TO_CHECK_SETTINGS.getName()) + .setNumberOfPartitions(TOPIC_TO_CHECK_SETTINGS.getNumberOfPartitions()) + .setMaxMessageBytes(TOPIC_TO_CHECK_SETTINGS.getMaxMessageBytes()) + .clickCreateTopicBtn(); + topicDetails + .waitUntilScreenReady(); + TOPIC_LIST.add(TOPIC_TO_CHECK_SETTINGS); + topicDetails + .openDetailsTab(SETTINGS); + topicSettingsTab + .waitUntilScreenReady(); + SoftAssertions softly = new SoftAssertions(); + softly.assertThat(topicSettingsTab.getValueByKey("retention.bytes")) + .as("getValueOfKey(retention.bytes)").isEqualTo(TOPIC_TO_CHECK_SETTINGS.getMaxSizeOnDisk().getOptionValue()); + softly.assertThat(topicSettingsTab.getValueByKey("max.message.bytes")) + .as("getValueOfKey(max.message.bytes)").isEqualTo(TOPIC_TO_CHECK_SETTINGS.getMaxMessageBytes()); + softly.assertAll(); + TOPIC_TO_CHECK_SETTINGS + .setMaxSizeOnDisk(SIZE_1_GB) + .setMaxMessageBytes("1000056"); + topicDetails + .openDotMenu() + .clickEditSettingsMenu(); + topicCreateEditForm + .waitUntilScreenReady() + .setMaxSizeOnDiskInGB(TOPIC_TO_CHECK_SETTINGS.getMaxSizeOnDisk()) + .setMaxMessageBytes(TOPIC_TO_CHECK_SETTINGS.getMaxMessageBytes()) + .clickCreateTopicBtn(); + topicDetails + .waitUntilScreenReady() + .openDetailsTab(SETTINGS); + topicSettingsTab + .waitUntilScreenReady(); + softly.assertThat(topicSettingsTab.getValueByKey("retention.bytes")) + .as("getValueOfKey(retention.bytes)").isEqualTo(TOPIC_TO_CHECK_SETTINGS.getMaxSizeOnDisk().getOptionValue()); + softly.assertThat(topicSettingsTab.getValueByKey("max.message.bytes")) + .as("getValueOfKey(max.message.bytes)").isEqualTo(TOPIC_TO_CHECK_SETTINGS.getMaxMessageBytes()); + softly.assertAll(); + } + @AfterAll public void afterAll() { TOPIC_LIST.forEach(topic -> apiHelper.deleteTopic(CLUSTER_NAME, topic.getName()));
test
val
2022-12-20T17:19:31
"2022-12-20T14:50:46Z"
ArthurNiedial
train
provectus/kafka-ui/3056_3122
provectus/kafka-ui
provectus/kafka-ui/3056
provectus/kafka-ui/3122
[ "connected" ]
5c723d9b4491019bc6de01e3746b5b86d70f0560
aead0e960c60f55ad7ca13285221addb966ed57b
[]
[]
"2022-12-23T05:46:48Z"
[ "scope/infrastructure" ]
Increment serde api version manually
[ ".github/workflows/release-serde-api.yaml" ]
[ ".github/workflows/release-serde-api.yaml" ]
[]
diff --git a/.github/workflows/release-serde-api.yaml b/.github/workflows/release-serde-api.yaml index e13fa7a23a0..227dbcec24c 100644 --- a/.github/workflows/release-serde-api.yaml +++ b/.github/workflows/release-serde-api.yaml @@ -27,7 +27,4 @@ jobs: - name: Publish to Maven Central run: | - MVN_VERSION=$(curl -s https://search.maven.org/solrsearch/select?q=g:"com.provectus"+AND+a:"kafka-ui-serde-api" | grep -o '"latestVersion": *"[^"]*"' | grep -o '"[^"]*"$' | sed 's/"//g') - MVN_VERSION=$(echo "$MVN_VERSION" | awk 'BEGIN{FS=OFS="."} {$2+=1} 1') - mvn -B -ntp versions:set -DnewVersion=$MVN_VERSION -pl kafka-ui-serde-api mvn source:jar javadoc:jar package gpg:sign -Dgpg.passphrase=${{ secrets.GPG_PASSPHRASE }} -Dserver.username=${{ secrets.NEXUS_USERNAME }} -Dserver.password=${{ secrets.NEXUS_PASSWORD }} nexus-staging:deploy -pl kafka-ui-serde-api -s settings.xml
null
train
val
2022-12-22T14:05:53
"2022-12-09T15:05:57Z"
Haarolean
train
provectus/kafka-ui/3120_3125
provectus/kafka-ui
provectus/kafka-ui/3120
provectus/kafka-ui/3125
[ "connected" ]
a9d7d6c5153ea63999f69f0b0433362a0b5a9c74
22439f30e803f4090489ded783055c0728bff6a2
[]
[ "we don't need to have several methods using same confirmBtn on the same place. remove clickConfirmRecreateBtn and reuse clickConfirmDeleteBtn pls. and rename the last is abstract way -> clickConfirmBtnMdl", "Fixed." ]
"2022-12-23T12:39:01Z"
[ "scope/QA", "scope/AQA" ]
[e2e] TopicTests.recreateTopicFromTopicProfile : Recreate topic from topic profile
Autotest implementation for: https://app.qase.io/case/KAFKAUI-247 Description: Not set Steps: 1)Navigate to Topics 2)Select the Topic 3)Select "Recreate Topic" from 3dot menu 4)Press "Confirm" button Expected results: 3)Confirm the action" pop up should appear 4)Success message should be displayed and Topic should be recreated
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java" ]
[ "kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java" ]
[ "kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java" ]
diff --git a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java index 6b328ab22e0..70d5934d048 100644 --- a/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java +++ b/kafka-ui-e2e-checks/src/main/java/com/provectus/kafka/ui/pages/topic/TopicDetails.java @@ -19,6 +19,7 @@ public class TopicDetails extends BasePage { protected SelenideElement clearMessagesBtn = $x(("//div[contains(text(), 'Clear messages')]")); + protected SelenideElement recreateTopicBtn = $x("//div[text()='Recreate Topic']"); protected SelenideElement messageAmountCell = $x("//tbody/tr/td[5]"); protected SelenideElement overviewTab = $x("//a[contains(text(),'Overview')]"); protected SelenideElement messagesTab = $x("//a[contains(text(),'Messages')]"); @@ -41,6 +42,7 @@ public class TopicDetails extends BasePage { protected SelenideElement cleanUpPolicyField = $x("//div[contains(text(),'Clean Up Policy')]/../span/*"); protected SelenideElement partitionsField = $x("//div[contains(text(),'Partitions')]/../span"); protected SelenideElement backToCreateFiltersLink = $x("//div[text()='Back To create filters']"); + protected SelenideElement confirmationMdl = $x("//div[text()= 'Confirm the action']/.."); protected ElementsCollection messageGridItems = $$x("//tbody//tr"); protected String seekFilterDdlLocator = "//ul[@id='selectSeekType']/ul/li[text()='%s']"; protected String savedFilterNameLocator = "//div[@role='savedFilter']/div[contains(text(),'%s')]"; @@ -85,12 +87,23 @@ public TopicDetails clickEditSettingsMenu() { return this; } + @Step + public boolean isConfirmationMdlVisible(){ + return isVisible(confirmationMdl); + } + @Step public TopicDetails clickClearMessagesMenu() { clearMessagesBtn.shouldBe(Condition.visible).click(); return this; } + @Step + public TopicDetails clickRecreateTopicMenu(){ + recreateTopicBtn.shouldBe(Condition.visible).click(); + return this; + } + @Step public String getCleanUpPolicy() { return cleanUpPolicyField.getText(); @@ -113,7 +126,7 @@ public TopicDetails clickDeleteTopicMenu() { } @Step - public TopicDetails clickConfirmDeleteBtn() { + public TopicDetails clickConfirmBtnMdl() { confirmBtn.shouldBe(Condition.enabled).click(); confirmBtn.shouldBe(Condition.disappear); return this;
diff --git a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java index 4eca498532f..fdeb95b19bf 100644 --- a/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java +++ b/kafka-ui-e2e-checks/src/test/java/com/provectus/kafka/ui/suite/topics/TopicsTests.java @@ -1,5 +1,6 @@ package com.provectus.kafka.ui.suite.topics; +import static com.provectus.kafka.ui.pages.BasePage.AlertHeader.SUCCESS; import static com.provectus.kafka.ui.pages.topic.TopicDetails.TopicMenu.MESSAGES; import static com.provectus.kafka.ui.pages.topic.TopicDetails.TopicMenu.SETTINGS; import static com.provectus.kafka.ui.pages.topic.enums.CleanupPolicyValue.COMPACT; @@ -169,7 +170,7 @@ public void deleteTopic() { topicDetails .openDotMenu() .clickDeleteTopicMenu() - .clickConfirmDeleteBtn(); + .clickConfirmBtnMdl(); navigateToTopics(); Assertions.assertFalse(topicsList.isTopicVisible(TOPIC_FOR_DELETE.getName()), "isTopicVisible"); TOPIC_LIST.remove(TOPIC_FOR_DELETE); @@ -439,6 +440,37 @@ void checkRetentionBytesAccordingToMaxSizeOnDisk(){ softly.assertAll(); } + @DisplayName("TopicTests.recreateTopicFromTopicProfile : Recreate topic from topic profile") + @Suite(suiteId = SUITE_ID, title = SUITE_TITLE) + @AutomationStatus(status = Status.AUTOMATED) + @CaseId(247) + @Test + void recreateTopicFromTopicProfile(){ + Topic topicToRecreate = new Topic() + .setName("topic-to-recreate-" + randomAlphabetic(5)) + .setNumberOfPartitions(1); + navigateToTopics(); + topicsList + .clickAddTopicBtn(); + topicCreateEditForm + .waitUntilScreenReady() + .setTopicName(topicToRecreate.getName()) + .setNumberOfPartitions(topicToRecreate.getNumberOfPartitions()) + .clickCreateTopicBtn(); + topicDetails + .waitUntilScreenReady(); + TOPIC_LIST.add(topicToRecreate); + topicDetails + .openDotMenu() + .clickRecreateTopicMenu(); + assertThat(topicDetails.isConfirmationMdlVisible()).as("isConfirmationMdlVisible()").isTrue(); + topicDetails + .clickConfirmBtnMdl(); + assertThat(topicDetails.isAlertWithMessageVisible(SUCCESS, + String.format("Topic %s successfully recreated!", topicToRecreate.getName()))) + .as("isAlertWithMessageVisible()").isTrue(); + } + @AfterAll public void afterAll() { TOPIC_LIST.forEach(topic -> apiHelper.deleteTopic(CLUSTER_NAME, topic.getName()));
test
val
2022-12-22T17:05:38
"2022-12-22T13:01:34Z"
ArthurNiedial
train
provectus/kafka-ui/3115_3145
provectus/kafka-ui
provectus/kafka-ui/3115
provectus/kafka-ui/3145
[ "connected" ]
566dab078f7e4f75804f0aa6ba706cc8d73a127e
aeda502b09d1dcf132f825d4b0779972037f74e9
[ "Hello there wanisfahmyDE! πŸ‘‹\n\nThank you and congratulations πŸŽ‰ for opening your very first issue in this project! πŸ’–\n\nIn case you want to claim this issue, please comment down below! We will try to get back to you as soon as we can. πŸ‘€", "@Mgrdich we can skip this param if all the partitions are selected\r\n\r\ncc: @iliax ", "Needs a fix.\r\n\r\n#3568 \r\nhttps://github.com/provectus/kafka-ui/pull/3582#discussion_r1161564606\r\n#3525" ]
[]
"2022-12-26T11:53:42Z"
[ "type/bug", "scope/frontend", "status/accepted", "status/confirmed" ]
HTTP 414 with a big partitions list selected
<!-- Don't forget to check for existing issues/discussions regarding your proposal. We might already have it. https://github.com/provectus/kafka-ui/issues https://github.com/provectus/kafka-ui/discussions --> **Describe the bug** running on latest master: messages do not load under messages tab for certain topics in a given cluster. all settings are default (serde, pagination). The response in the browser console is 414 (url too long). The problematic endpoint is `t/opics/mainOsyTrigger/messages?filterQueryType` notes: * This kafka cluster has 240 partitions, maybe that's the problem. * This problem doesn't occur on release `v0.4.0` The request URL looks like this: `https:/my-kafka-ui/topics/mainOsyTrigger/messages?filterQueryType=STRING_CONTAINS&attempt=1&limit=100&seekDirection=FORWARD&keySerde=null&valueSerde=null&seekType=BEGINNING&seekTo=0%3A%3A3074090%2C1%3A%3A2873466%2C2%3A%3A3057371%2C3%3A%3A3224533%2C4%3A%3A3238047%2C5%3A%3A2882461%2C6%3A%3A2973712%2C7%3A%3A3036851%2C8%3A%3A2998608%2C9%3A%3A3081134%2C10%3A%3A3432678%2C11%3A%3A3032894%2C12%3A%3A2932963%2C13%3A%3A3021701%2C14%3A%3A3093426%2C15%3A%3A3142236%2C16%3A%3A3346479%2C17%3A%3A3042457%2C18%3A%3A3188474%2C19%3A%3A3122717%2C20%3A%3A3338328%2C21%3A%3A3002992%2C22%3A%3A3117563%2C23%3A%3A3066825%2C24%3A%3A3024334%2C25%3A%3A3090044%2C26%3A%3A2875326%2C27%3A%3A3071587%2C28%3A%3A3073475%2C29%3A%3A3353610%2C30%3A%3A3060803%2C31%3A%3A2985938%2C32%3A%3A3300810%2C33%3A%3A3043026%2C34%3A%3A3088143%2C35%3A%3A3094591%2C36%3A%3A3148360%2C37%3A%3A2972083%2C38%3A%3A3222534%2C39%3A%3A3014068%2C40%3A%3A3102879%2C41%3A%3A2964547%2C42%3A%3A2936218%2C43%3A%3A2996908%2C44%3A%3A3061150%2C45%3A%3A2790067%2C46%3A%3A2970756%2C47%3A%3A3039456%2C48%3A%3A3097037%2C49%3A%3A3156340%2C50%3A%3A3238510%2C51%3A%3A2962902%2C52%3A%3A3066628%2C53%3A%3A2975306%2C54%3A%3A3180922%2C55%3A%3A3410678%2C56%3A%3A2992099%2C57%3A%3A3155103%2C58%3A%3A2938148%2C59%3A%3A2929215%2C60%3A%3A3142161%2C61%3A%3A3202679%2C62%3A%3A3503614%2C63%3A%3A3113890%2C64%3A%3A2980539%2C65%3A%3A3055035%2C66%3A%3A2922145%2C67%3A%3A3094968%2C68%3A%3A3229233%2C69%3A%3A3032187%2C70%3A%3A2960532%2C71%3A%3A3200119%2C72%3A%3A3072256%2C73%3A%3A3336365%2C74%3A%3A3071080%2C75%3A%3A3091609%2C76%3A%3A3012113%2C77%3A%3A3096383%2C78%3A%3A2929622%2C79%3A%3A3095415%2C80%3A%3A3043199%2C81%3A%3A3283358%2C82%3A%3A3056106%2C83%3A%3A2983369%2C84%3A%3A2916423%2C85%3A%3A3275715%2C86%3A%3A3296668%2C87%3A%3A3394228%2C88%3A%3A3098087%2C89%3A%3A3030158%2C90%3A%3A3150636%2C91%3A%3A2878705%2C92%3A%3A3039426%2C93%3A%3A3150403%2C94%3A%3A3048218%2C95%3A%3A3001379%2C96%3A%3A3163979%2C97%3A%3A3299770%2C98%3A%3A3071348%2C99%3A%3A3103185%2C100%3A%3A3078464%2C101%3A%3A3136160%2C102%3A%3A3218327%2C103%3A%3A2886828%2C104%3A%3A3032169%2C105%3A%3A3003752%2C106%3A%3A2896270%2C107%3A%3A3160299%2C108%3A%3A3103092%2C109%3A%3A3107543%2C110%3A%3A2892965%2C111%3A%3A3158000%2C112%3A%3A2940465%2C113%3A%3A3019562%2C114%3A%3A3383091%2C115%3A%3A3052166%2C116%3A%3A3092634%2C117%3A%3A3461752%2C118%3A%3A3173274%2C119%3A%3A3071851%2C120%3A%3A3106930%2C121%3A%3A3111206%2C122%3A%3A3060124%2C123%3A%3A3156639%2C124%3A%3A3130224%2C125%3A%3A2902996%2C126%3A%3A2973758%2C127%3A%3A3292917%2C128%3A%3A3021322%2C129%3A%3A3178655%2C130%3A%3A3142298%2C131%3A%3A2952581%2C132%3A%3A3099713%2C133%3A%3A3097531%2C134%3A%3A2949911%2C135%3A%3A3120310%2C136%3A%3A2943906%2C137%3A%3A3081907%2C138%3A%3A3197362%2C139%3A%3A2958921%2C140%3A%3A3022182%2C141%3A%3A3129833%2C142%3A%3A2963757%2C143%3A%3A3223688%2C144%3A%3A2991264%2C145%3A%3A3021824%2C146%3A%3A3309025%2C147%3A%3A3281467%2C148%3A%3A3032307%2C149%3A%3A2872527%2C150%3A%3A2996964%2C151%3A%3A3036588%2C152%3A%3A3334637%2C153%3A%3A3346481%2C154%3A%3A3095406%2C155%3A%3A3035255%2C156%3A%3A3162874%2C157%3A%3A3153442%2C158%3A%3A3084418%2C159%3A%3A3212934%2C160%3A%3A2866368%2C161%3A%3A2918106%2C162%3A%3A3122933%2C163%3A%3A3372647%2C164%3A%3A3493517%2C165%3A%3A3229364%2C166%3A%3A3084963%2C167%3A%3A3076450%2C168%3A%3A3126261%2C169%3A%3A3132226%2C170%3A%3A2989424%2C171%3A%3A2955585%2C172%3A%3A2980134%2C173%3A%3A2991709%2C174%3A%3A3436223%2C175%3A%3A3059177%2C176%3A%3A2949204%2C177%3A%3A3060454%2C178%3A%3A3067493%2C179%3A%3A3053507%2C180%3A%3A3080837%2C181%3A%3A3068181%2C182%3A%3A3080689%2C183%3A%3A3120516%2C184%3A%3A3202197%2C185%3A%3A3049898%2C186%3A%3A3150783%2C187%3A%3A3093330%2C188%3A%3A3168716%2C189%3A%3A2981076%2C190%3A%3A2897306%2C191%3A%3A3045436%2C192%3A%3A3114719%2C193%3A%3A3153424%2C194%3A%3A3338991%2C195%3A%3A3001310%2C196%3A%3A3163588%2C197%3A%3A3143609%2C198%3A%3A3321556%2C199%3A%3A3096641%2C200%3A%3A3138570%2C201%3A%3A3305132%2C202%3A%3A3083234%2C203%3A%3A2866230%2C204%3A%3A3239104%2C205%3A%3A2932237%2C206%3A%3A3026534%2C207%3A%3A3180500%2C208%3A%3A3246722%2C209%3A%3A3025870%2C210%3A%3A3090457%2C211%3A%3A3122387%2C212%3A%3A2812894%2C213%3A%3A3111089%2C214%3A%3A3255140%2C215%3A%3A3064275%2C216%3A%3A3239711%2C217%3A%3A2770812%2C218%3A%3A3160334%2C219%3A%3A3347422%2C220%3A%3A3116640%2C221%3A%3A3239702%2C222%3A%3A2832341%2C223%3A%3A3096067%2C224%3A%3A3371181%2C225%3A%3A3328135%2C226%3A%3A3220448%2C227%3A%3A3035209%2C228%3A%3A3096710%2C229%3A%3A3017020%2C230%3A%3A3054040%2C231%3A%3A3075303%2C232%3A%3A2964717%2C233%3A%3A2915869%2C234%3A%3A3099099%2C235%3A%3A3019695%2C236%3A%3A2826167%2C237%3A%3A3046179%2C238%3A%3A3065352%2C239%3A%3A3045407` **Set up** deploying latest master branch using a k8s deployment manifest, passing configs as env variables. **Steps to Reproduce** Steps to reproduce the behavior: 1. open kafka-ui from browser 2. navigate to cluster --> topics --> meesages tab 3. nothing is displayed, error in browser console (described above) **Expected behavior** messages are loaded correctly
[ "kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx" ]
[ "kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx" ]
[]
diff --git a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx index 096a714c4ee..08e4aed6912 100644 --- a/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx +++ b/kafka-ui-react-app/src/components/Topics/Topic/Messages/Filters/Filters.tsx @@ -231,18 +231,21 @@ const Filters: React.FC<FiltersProps> = ({ props.seekType = SeekType.TIMESTAMP; } - props.seekTo = selectedPartitions.map(({ value }) => { - const offsetProperty = - seekDirection === SeekDirection.FORWARD ? 'offsetMin' : 'offsetMax'; - const offsetBasedSeekTo = - currentOffset || partitionMap[value][offsetProperty]; - const seekToOffset = - currentSeekType === SeekType.OFFSET - ? offsetBasedSeekTo - : timestamp?.getTime(); - - return `${value}::${seekToOffset || '0'}`; - }); + if (selectedPartitions.length !== partitions.length) { + // not everything in the partition is selected + props.seekTo = selectedPartitions.map(({ value }) => { + const offsetProperty = + seekDirection === SeekDirection.FORWARD ? 'offsetMin' : 'offsetMax'; + const offsetBasedSeekTo = + currentOffset || partitionMap[value][offsetProperty]; + const seekToOffset = + currentSeekType === SeekType.OFFSET + ? offsetBasedSeekTo + : timestamp?.getTime(); + + return `${value}::${seekToOffset || '0'}`; + }); + } } const newProps = omitBy(props, (v) => v === undefined || v === '');
null
train
val
2023-01-12T09:20:54
"2022-12-21T13:02:28Z"
wanisfahmyDE
train
provectus/kafka-ui/3154_3155
provectus/kafka-ui
provectus/kafka-ui/3154
provectus/kafka-ui/3155
[ "keyword_pr_to_issue" ]
5f4bb43afd3a516176db481944fb5adb3178a40e
8b126d843ee17dc2b79590d951d9b28e8e85f3ca
[]
[]
"2022-12-27T19:40:19Z"
[ "type/bug", "scope/backend", "status/accepted", "status/confirmed", "area/rbac" ]
Google provider w/ OAuth2 w/ RBAC: Invalid client registration id
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/config/auth/OAuthPropertiesConverter.java" ]
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/config/auth/OAuthPropertiesConverter.java" ]
[]
diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/config/auth/OAuthPropertiesConverter.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/config/auth/OAuthPropertiesConverter.java index 8fffbfb6997..c3d20664914 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/config/auth/OAuthPropertiesConverter.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/config/auth/OAuthPropertiesConverter.java @@ -8,12 +8,14 @@ import lombok.NoArgsConstructor; import org.apache.commons.lang3.StringUtils; import org.springframework.boot.autoconfigure.security.oauth2.client.OAuth2ClientProperties; +import org.springframework.security.config.oauth2.client.CommonOAuth2Provider; @NoArgsConstructor(access = AccessLevel.PRIVATE) public final class OAuthPropertiesConverter { private static final String TYPE = "type"; private static final String GOOGLE = "google"; + public static final String DUMMY = "dummy"; public static OAuth2ClientProperties convertProperties(final OAuthProperties properties) { final var result = new OAuth2ClientProperties(); @@ -57,7 +59,14 @@ private static void applyGoogleTransformations(OAuth2Provider provider) { return; } - final String newUri = provider.getAuthorizationUri() + "?hd=" + allowedDomain; + String authorizationUri = CommonOAuth2Provider.GOOGLE + .getBuilder(DUMMY) + .clientId(DUMMY) + .build() + .getProviderDetails() + .getAuthorizationUri(); + + final String newUri = authorizationUri + "?hd=" + allowedDomain; provider.setAuthorizationUri(newUri); }
null
val
val
2022-12-27T09:45:17
"2022-12-27T19:39:00Z"
Haarolean
train
provectus/kafka-ui/3141_3157
provectus/kafka-ui
provectus/kafka-ui/3141
provectus/kafka-ui/3157
[ "connected" ]
16750bd899443dd696e03e2c2254fe2973d0eb20
0cba8240b7d6aff7f9417c5651bf811f935254b3
[ "@fisherwei hi, please share a full stacktrace (you've omitted the part below)", "Same here. Full stacktrace included, after this it just repeats.\r\n\r\n```provectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: java.lang.IllegalStateException: Error while creating AdminClient for Cluster XXX\r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #011at com.provectus.kafka.ui.service.AdminClientServiceImpl.lambda$createAdminClient$3(AdminClientServiceImpl.java:45)\r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #011at reactor.core.publisher.Mono.lambda$onErrorMap$31(Mono.java:3776)\r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #011at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onError(FluxOnErrorResume.java:94)\r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #011at reactor.core.publisher.FluxHide$SuppressFuseableSubscriber.onError(FluxHide.java:142)\r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #011at reactor.core.publisher.FluxMap$MapSubscriber.onError(FluxMap.java:134)\r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #011at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.onError(Operators.java:2063)\r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #011at reactor.core.publisher.MonoFlatMap$FlatMapMain.onError(MonoFlatMap.java:172)\r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #011at reactor.core.publisher.MonoFlatMap$FlatMapMain.onError(MonoFlatMap.java:172)\r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #011at reactor.core.publisher.MonoFlatMap$FlatMapMain.onError(MonoFlatMap.java:172)\r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #011at reactor.core.publisher.MonoIgnoreThen$ThenIgnoreMain.onError(MonoIgnoreThen.java:278)\r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #011at reactor.core.publisher.MonoPublishOn$PublishOnSubscriber.run(MonoPublishOn.java:187)\r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #011at reactor.core.scheduler.SchedulerTask.call(SchedulerTask.java:68)\r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #011at reactor.core.scheduler.SchedulerTask.call(SchedulerTask.java:28)\r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #011at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)\r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #011at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)\r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #011at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #011at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #011at java.base/java.lang.Thread.run(Thread.java:833)\r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: Caused by: org.apache.kafka.common.errors.UnsupportedVersionException: Attempted to write a non-default includeClusterAuthorizedOperations at version 5\r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #033[30m2022-12-27 15:18:36,575#033[0;39m #033[39mDEBUG#033[0;39m [#033[34mparallel-11#033[0;39m] #033[33mc.p.k.u.s.ClustersStatisticsScheduler#033[0;39m: Metrics updated for cluster: XXX\r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #033[30m2022-12-27 15:18:36,592#033[0;39m #033[34mINFO #033[0;39m [#033[34mkafka-admin-client-thread | kafka-ui-admin-client-1672153416196#033[0;39m] #033[33mo.a.k.c.NetworkClient#033[0;39m: [AdminClient clientId=kafka-ui-admin-client-1672153416196] Node 1004 \r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #033[30m2022-12-27 15:18:36,640#033[0;39m #033[34mINFO #033[0;39m [#033[34mkafka-admin-client-thread | kafka-ui-admin-client-1672153716196#033[0;39m] #033[33mo.a.k.c.NetworkClient#033[0;39m: [AdminClient clientId=kafka-ui-admin-client-1672153716196] Node 1 \r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #033[30m2022-12-27 15:18:36,697#033[0;39m #033[34mINFO #033[0;39m [#033[34mkafka-admin-client-thread | kafka-ui-admin-client-1672153116196#033[0;39m] #033[33mo.a.k.c.NetworkClient#033[0;39m: [AdminClient clientId=kafka-ui-admin-client-1672153116196] Node 1004 \r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #033[30m2022-12-27 15:18:36,897#033[0;39m #033[34mINFO #033[0;39m [#033[34mkafka-admin-client-thread | kafka-ui-admin-client-1672153416196#033[0;39m] #033[33mo.a.k.c.NetworkClient#033[0;39m: [AdminClient clientId=kafka-ui-admin-client-1672153416196] Node 2 \r\nprovectuslabs/kafka-ui:v0.5.0|c94d85724408[2661]: #033[30m2022-12-27 15:18:36,965#033[0;39m #033[34mINFO #033[0;39m [#033[34mkafka-admin-client-thread | kafka-ui-admin-client-1672152516196#033[0;39m] #033[33mo.a.k.c.NetworkClient#033[0;39m: [AdminClient clientId=kafka-ui-admin-client-1672152516196] Node 1002 \r\n```\r\n\r\n", "Thanks. It seems like your kafka doesn't support ACLs operations we do, we'll take a look for a workaround.", "I had this issue. The issue is that your kafka version does not support includeAuthorizedOperations(true). A work around is setting this to false\r\n\r\nIn this method [https://github.com/provectus/kafka-ui/blob/master/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java#L306](url) change it to false and it will work ", "Hello @fisherwei , thank you for ticket. \r\nCan you please check if it was fixed with `master` tag? If so, we will add fix this to 0.5.x bugfix release.", "I didn't open the ticket but had the same problem. `master` works for us πŸ‘ \r\n\r\nThanks! πŸ™ ", "@RoccoMathijn , thank you for confirmation", "When do you plan to push an official 0.5.x patch image with this fix?", "@kewitschka you can pull master image for now" ]
[]
"2022-12-28T08:54:33Z"
[ "type/bug", "scope/backend", "status/accepted", "status/confirmed" ]
[BE] Bug: UnsupportedVersionException: includeClusterAuthorizedOperations
application.yml ``` kafka: clusters: - name: qsmzq-prod bootstrapServers: node1:9092,node2:9092,node3:9092 properties: security.protocol: SASL_PLAINTEXT sasl.mechanism: SCRAM-SHA-256 sasl.jaas.config: 'org.apache.kafka.common.security.scram.ScramLoginModule required username="kafka" password="password";' spring: jmx: enabled: false auth: type: DISABLED ``` ``` java.lang.IllegalStateException: Error while creating AdminClient for Cluster qsmzq-prod at com.provectus.kafka.ui.service.AdminClientServiceImpl.lambda$createAdminClient$3(AdminClientServiceImpl.java:45) at reactor.core.publisher.Mono.lambda$onErrorMap$31(Mono.java:3776) at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onError(FluxOnErrorResume.java:94) at reactor.core.publisher.FluxHide$SuppressFuseableSubscriber.onError(FluxHide.java:142) at reactor.core.publisher.FluxMap$MapSubscriber.onError(FluxMap.java:134) at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.onError(Operators.java:2063) at reactor.core.publisher.MonoFlatMap$FlatMapMain.onError(MonoFlatMap.java:172) at reactor.core.publisher.MonoFlatMap$FlatMapMain.onError(MonoFlatMap.java:172) at reactor.core.publisher.MonoFlatMap$FlatMapMain.onError(MonoFlatMap.java:172) at reactor.core.publisher.MonoIgnoreThen$ThenIgnoreMain.onError(MonoIgnoreThen.java:278) at reactor.core.publisher.MonoPublishOn$PublishOnSubscriber.run(MonoPublishOn.java:187) at reactor.core.scheduler.SchedulerTask.call(SchedulerTask.java:68) at reactor.core.scheduler.SchedulerTask.call(SchedulerTask.java:28) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:833) Caused by: org.apache.kafka.common.errors.UnsupportedVersionException: Attempted to write a non-default includeClusterAuthorizedOperations at version 7 ``` I used same config at v0.4.0, everything works as expected.
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java" ]
[ "kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java" ]
[]
diff --git a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java index b06716e6314..1ffbd429180 100644 --- a/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java +++ b/kafka-ui-api/src/main/java/com/provectus/kafka/ui/service/ReactiveAdminClient.java @@ -43,6 +43,7 @@ import org.apache.kafka.clients.admin.ConsumerGroupDescription; import org.apache.kafka.clients.admin.ConsumerGroupListing; import org.apache.kafka.clients.admin.DescribeClusterOptions; +import org.apache.kafka.clients.admin.DescribeClusterResult; import org.apache.kafka.clients.admin.DescribeConfigsOptions; import org.apache.kafka.clients.admin.ListConsumerGroupOffsetsOptions; import org.apache.kafka.clients.admin.ListOffsetsResult; @@ -80,7 +81,8 @@ public class ReactiveAdminClient implements Closeable { private enum SupportedFeature { INCREMENTAL_ALTER_CONFIGS(2.3f), - CONFIG_DOCUMENTATION_RETRIEVAL(2.6f); + CONFIG_DOCUMENTATION_RETRIEVAL(2.6f), + DESCRIBE_CLUSTER_INCLUDE_AUTHORIZED_OPERATIONS(2.3f); private final float sinceVersion; @@ -300,11 +302,14 @@ public Mono<Map<Integer, Map<String, DescribeLogDirsResponse.LogDirInfo>>> descr } public Mono<ClusterDescription> describeCluster() { - return describeClusterImpl(client); + return describeClusterImpl(client, features); } - private static Mono<ClusterDescription> describeClusterImpl(AdminClient client) { - var result = client.describeCluster(new DescribeClusterOptions().includeAuthorizedOperations(true)); + private static Mono<ClusterDescription> describeClusterImpl(AdminClient client, Set<SupportedFeature> features) { + boolean includeAuthorizedOperations = + features.contains(SupportedFeature.DESCRIBE_CLUSTER_INCLUDE_AUTHORIZED_OPERATIONS); + DescribeClusterResult result = client.describeCluster( + new DescribeClusterOptions().includeAuthorizedOperations(includeAuthorizedOperations)); var allOfFuture = KafkaFuture.allOf( result.controller(), result.clusterId(), result.nodes(), result.authorizedOperations()); return toMono(allOfFuture).then( @@ -320,7 +325,7 @@ private static Mono<ClusterDescription> describeClusterImpl(AdminClient client) } private static Mono<String> getClusterVersion(AdminClient client) { - return describeClusterImpl(client) + return describeClusterImpl(client, Set.of()) // choosing node from which we will get configs (starting with controller) .flatMap(descr -> descr.controller != null ? Mono.just(descr.controller)
null
val
val
2022-12-28T15:39:52
"2022-12-26T07:47:23Z"
fisherwei
train