From patchwork Fri Mar 18 13:51:32 2016 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: "H.J. Lu" X-Patchwork-Id: 599465 Return-Path: X-Original-To: incoming@patchwork.ozlabs.org Delivered-To: patchwork-incoming@bilbo.ozlabs.org Received: from sourceware.org (server1.sourceware.org [209.132.180.131]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by ozlabs.org (Postfix) with ESMTPS id 3qRRV13Fy9z9sD5 for ; Sat, 19 Mar 2016 00:51:53 +1100 (AEDT) Authentication-Results: ozlabs.org; dkim=pass (1024-bit key; secure) header.d=sourceware.org header.i=@sourceware.org header.b=VF18vJaT; dkim-atps=neutral DomainKey-Signature: a=rsa-sha1; c=nofws; d=sourceware.org; h=list-id :list-unsubscribe:list-subscribe:list-archive:list-post :list-help:sender:mime-version:in-reply-to:references:date :message-id:subject:from:to:cc:content-type; q=dns; s=default; b= ruXeq00R+q4iEe1kSMfcL2mXczM4PxfgLk+jLYnG8YY7wPE4+NzOFdLyIlRWjBp3 iXuUMfU+QkFxXHeNQnStJ/zFoSPSmT1BJc0V31m4Ovx8MXLC60t+4HNy8AQ6NN5d v2XjqW1ZHx5PGnnSVfyT3SXzSB2Bic9ZWo/6etUTgaY= DKIM-Signature: v=1; a=rsa-sha1; c=relaxed; d=sourceware.org; h=list-id :list-unsubscribe:list-subscribe:list-archive:list-post :list-help:sender:mime-version:in-reply-to:references:date :message-id:subject:from:to:cc:content-type; s=default; bh=jxg0v bbM+VONEI8gspkfK4DaxcU=; b=VF18vJaTSq8gcVzF51ypLElPDJmX87wDbwujZ wS1G+4G6h3dOXc4ZV7Q3UEIi1ZQHJLQY+sIvhGBuEMx+WVHIBCDsiqq3YHAFf6n5 eIPC/WxmZMpSRciIjQAyw2TMbHdJohatL5wqFnB4us9ZYTi3uFSbZxbLtB6D4Ihv 4UoOro= Received: (qmail 103948 invoked by alias); 18 Mar 2016 13:51:47 -0000 Mailing-List: contact libc-alpha-help@sourceware.org; run by ezmlm Precedence: bulk List-Id: List-Unsubscribe: List-Subscribe: List-Archive: List-Post: List-Help: , Sender: libc-alpha-owner@sourceware.org Delivered-To: mailing list libc-alpha@sourceware.org Received: (qmail 103924 invoked by uid 89); 18 Mar 2016 13:51:46 -0000 Authentication-Results: sourceware.org; auth=none X-Virus-Found: No X-Spam-SWARE-Status: No, score=-2.4 required=5.0 tests=AWL, BAYES_00, FREEMAIL_FROM, RCVD_IN_DNSWL_LOW, SPF_PASS autolearn=ham version=3.3.2 spammy=429, AmitPawaramdcom, Amit.Pawar@amd.com, amitpawaramdcom X-HELO: mail-qg0-f45.google.com X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:mime-version:in-reply-to:references:date :message-id:subject:from:to:cc; bh=MdsZgdWhAyhvLGYJD3RI/V/vFz2EKPi2P53EJnP8q1A=; b=HjJT3Z0rdYeK75p+9VVslcUHGL0u2MGIjFjipfofawP5M9U36mRnQyAvpviaj6AoNf la806M39fTXFCpRS5o5XOdMGNHmm6TfCTCBBBXuVfV7sCHG/Qgdo/iVaW1Rd6u9Fie9g Z99PiMrzXcTIWmYW40aj6Fizd823LC9RJyJHxiAIjdiHn/SufjauZGQ+EMay/2VJN0Ge /Ik2k4fjtHx6OSSiX0kjfvSOqNwNFD2zWaAYO9iNtqe/4ULlOp2CItY8S4My4ehqO6Lk QatZCJrsi6ltXPiAp6BWLTqg18uu7K2E04pg43C7tyCshMchqLGMKG9nEqTj0Vw1GYxr sy+Q== X-Gm-Message-State: AD7BkJIxWKjx3orCw40kUnPHkgWt2moeuNcCmlYbN33zqf6PgKly54AWHrTZSQLTuM2tcfUE7BgLYl3ycKiEcg== MIME-Version: 1.0 X-Received: by 10.141.1.87 with SMTP id c84mr23538822qhd.1.1458309092980; Fri, 18 Mar 2016 06:51:32 -0700 (PDT) In-Reply-To: References: Date: Fri, 18 Mar 2016 06:51:32 -0700 Message-ID: Subject: Re: [PATCH x86_64] Update memcpy, mempcpy and memmove selection order for Excavator CPU BZ #19583 From: "H.J. Lu" To: "Pawar, Amit" Cc: "libc-alpha@sourceware.org" On Fri, Mar 18, 2016 at 6:22 AM, Pawar, Amit wrote: >>No, it isn't fixed. Avoid_AVX_Fast_Unaligned_Load should disable __memcpy_avx_unaligned and nothing more. Also you need to fix ALL selections. > > diff --git a/sysdeps/x86_64/multiarch/memcpy.S b/sysdeps/x86_64/multiarch/memcpy.S > index 8882590..a5afaf4 100644 > --- a/sysdeps/x86_64/multiarch/memcpy.S > +++ b/sysdeps/x86_64/multiarch/memcpy.S > @@ -39,6 +39,8 @@ ENTRY(__new_memcpy) > ret > #endif > 1: lea __memcpy_avx_unaligned(%rip), %RAX_LP > + HAS_ARCH_FEATURE (Avoid_AVX_Fast_Unaligned_Load) > + jnz 3f > HAS_ARCH_FEATURE (AVX_Fast_Unaligned_Load) > jnz 2f > lea __memcpy_sse2_unaligned(%rip), %RAX_LP > @@ -52,6 +54,8 @@ ENTRY(__new_memcpy) > jnz 2f > lea __memcpy_ssse3(%rip), %RAX_LP > 2: ret > +3: lea __memcpy_ssse3(%rip), %RAX_LP > + ret > END(__new_memcpy) > > # undef ENTRY > > Will update all IFUNC's if this ok else please suggest. > Better, but not OK. Try something like iff --git a/sysdeps/x86_64/multiarch/memcpy.S b/sysdeps/x86_64/multiarch/memcpy.S index ab5998c..2abe2fd 100644 --- a/sysdeps/x86_64/multiarch/memcpy.S +++ b/sysdeps/x86_64/multiarch/memcpy.S @@ -42,9 +42,11 @@ ENTRY(__new_memcpy) ret #endif 1: lea __memcpy_avx_unaligned(%rip), %RAX_LP + HAS_ARCH_FEATURE (Avoid_AVX_Fast_Unaligned_Load) + jnz 3f HAS_ARCH_FEATURE (AVX_Fast_Unaligned_Load) jnz 2f - lea __memcpy_sse2_unaligned(%rip), %RAX_LP +3: lea __memcpy_sse2_unaligned(%rip), %RAX_LP HAS_ARCH_FEATURE (Fast_Unaligned_Load) jnz 2f lea __memcpy_sse2(%rip), %RAX_LP